Advances in Human Factors in Cybersecurity: Tareq Ahram Waldemar Karwowski Editors
Advances in Human Factors in Cybersecurity: Tareq Ahram Waldemar Karwowski Editors
Tareq Ahram
Waldemar Karwowski Editors
Advances in
Human Factors
in Cybersecurity
Proceedings of the AHFE 2019
International Conference on Human
Factors in Cybersecurity, July 24–28,
2019, Washington D.C., USA
Advances in Intelligent Systems and Computing
Volume 960
Series Editor
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences,
Warsaw, Poland
Advisory Editors
Nikhil R. Pal, Indian Statistical Institute, Kolkata, India
Rafael Bello Perez, Faculty of Mathematics, Physics and Computing,
Universidad Central de Las Villas, Santa Clara, Cuba
Emilio S. Corchado, University of Salamanca, Salamanca, Spain
Hani Hagras, School of Computer Science & Electronic Engineering,
University of Essex, Colchester, UK
László T. Kóczy, Department of Automation, Széchenyi István University,
Gyor, Hungary
Vladik Kreinovich, Department of Computer Science, University of Texas
at El Paso, El Paso, TX, USA
Chin-Teng Lin, Department of Electrical Engineering, National Chiao
Tung University, Hsinchu, Taiwan
Jie Lu, Faculty of Engineering and Information Technology,
University of Technology Sydney, Sydney, NSW, Australia
Patricia Melin, Graduate Program of Computer Science, Tijuana Institute
of Technology, Tijuana, Mexico
Nadia Nedjah, Department of Electronics Engineering, University of Rio de Janeiro,
Rio de Janeiro, Brazil
Ngoc Thanh Nguyen, Faculty of Computer Science and Management,
Wrocław University of Technology, Wrocław, Poland
Jun Wang, Department of Mechanical and Automation Engineering,
The Chinese University of Hong Kong, Shatin, Hong Kong
The series “Advances in Intelligent Systems and Computing” contains publications
on theory, applications, and design methods of Intelligent Systems and Intelligent
Computing. Virtually all disciplines such as engineering, natural sciences, computer
and information science, ICT, economics, business, e-commerce, environment,
healthcare, life science are covered. The list of topics spans all the areas of modern
intelligent systems and computing such as: computational intelligence, soft comput-
ing including neural networks, fuzzy systems, evolutionary computing and the fusion
of these paradigms, social intelligence, ambient intelligence, computational neuro-
science, artificial life, virtual worlds and society, cognitive science and systems,
Perception and Vision, DNA and immune based systems, self-organizing and
adaptive systems, e-Learning and teaching, human-centered and human-centric
computing, recommender systems, intelligent control, robotics and mechatronics
including human-machine teaming, knowledge-based paradigms, learning para-
digms, machine ethics, intelligent data analysis, knowledge management, intelligent
agents, intelligent decision making and support, intelligent network security, trust
management, interactive entertainment, Web intelligence and multimedia.
The publications within “Advances in Intelligent Systems and Computing” are
primarily proceedings of important conferences, symposia and congresses. They
cover significant recent developments in the field, both of a foundational and
applicable character. An important characteristic feature of the series is the short
publication time and world-wide distribution. This permits a rapid and broad
dissemination of research results.
Editors
123
Editors
Tareq Ahram Waldemar Karwowski
Institute for Advanced Systems Engineering University of Central Florida
University of Central Florida Orlando, FL, USA
Orlando, FL, USA
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Advances in Human Factors
and Ergonomics 2019
10th International Conference on Applied Human Factors and Ergonomics and the
Affiliated Conferences
v
vi Advances in Human Factors and Ergonomics 2019
(continued)
Advances in Artificial Intelligence, Software Tareq Ahram
and Systems Engineering
Advances in Human Factors in Architecture, Jerzy Charytonowicz and Christianne
Sustainable Urban Planning and Infrastructure Falcão
Advances in Physical Ergonomics and Human Ravindra S. Goonetilleke and Waldemar
Factors Karwowski
Advances in Interdisciplinary Practice in Industrial Cliff Sungsoo Shin
Design
Advances in Safety Management and Human Pedro M. Arezes
Factors
Advances in Social and Occupational Ergonomics Richard H. M. Goossens and Atsuo
Murata
Advances in Manufacturing, Production Waldemar Karwowski, Stefan
Management and Process Control Trzcielinski and Beata Mrugalska
Advances in Usability and User Experience Tareq Ahram and Christianne Falcão
Advances in Human Factors in Wearable Tareq Ahram
Technologies and Game Design
Advances in Human Factors in Communication Amic G. Ho
of Design
Advances in Additive Manufacturing, Modeling Massimo Di Nicolantonio, Emilio Rossi
Systems and 3D Prototyping and Thomas Alexander
Preface
Our daily life, economic vitality, and national security depend on a stable, safe, and
resilient cyberspace. We rely on this vast array of networks to communicate and
travel, power our homes, run our economy, and provide government services. Yet,
cyber intrusions and attacks have increased dramatically over the last decade,
exposing sensitive personal and business information, disrupting critical operations,
and imposing high costs on the economy. The human factor at the core of
cybersecurity provides greater insight into this issue and highlights human error and
awareness as key factors, in addition to technical lapses, as the areas of greatest
concern. This book focuses on the social, economic, and behavioral aspects of
cyberspace, which are largely missing from the general discourse on cybersecurity.
The human element at the core of cybersecurity is what makes cyberspace the
complex, adaptive system that it is. An inclusive, multi-disciplinary, holistic
approach that combines the technical and behavioral element is needed to enhance
cybersecurity. Human factors also pervade the top cyber threats. Personnel
management and cyber awareness are essential for achieving holistic cybersecurity.
This book will be of special value to a large variety of professionals, researchers,
and students focusing on the human aspect of cyberspace, and for the effective
evaluation of security measures, interfaces, user-centered design, and design for
special populations, particularly the elderly. We hope this book is informative, but
even more than that it is thought-provoking. We hope it inspires, leading the reader
to contemplate other questions, applications, and potential solutions in creating safe
and secure designs for all.
This book includes two main sections:
Section 1 Cybersecurity Applications and Privacy Research
Section 2 Awareness and Cyber-Physical Security
vii
viii Preface
Each section contains research papers that have been reviewed by members
of the International Editorial Board. Our sincere thanks and appreciation to the
Board members as listed below:
Ritu Chadha, USA
Grit Denker, USA
Frank Greitzer, USA
Jim Jones, USA
Denise Nicholson, USA
Anne Tall, USA
Mike Ter Louw, USA
Elizabeth Whitaker, USA
ix
x Contents
1 Introduction
individual is, the more likely they are to process the content of pop-up messages using
automatic heuristic processing compared to someone who is more suspicious and less
trusting who will likely engage in more cognitively effortful and time-consuming
processing. Similarly, those who have a higher need for cognitive stimulation [19], will
be more susceptible to influence techniques used within pop-up messages such as
urgency, compliance with authority and avoidance of loss; at the expense of looking for
suspicious aspects, such as message authenticity cues (e.g., correct spelling and
grammar, company name). This leads to a prediction that an intervention training
protocol that increases suspicion and encourages more effortful processing of pop-up
message content should have carryover effects to a subsequent task performed with
malevolent pop-up interruptions.
To our knowledge, only two published studies have considered human suscepti-
bility to fraudulent pop-up interruptions occurring during a demanding memory-based
task. [2] developed a paradigm where young adult participants were interrupted by one
of three different types of pop-up message during a serial recall memory recall task.
One third of pop-ups were designed to look genuine (genuine condition) and high in
authority with no cues to potential malevolence. Another third (mimicked condition)
were also high in authority but contained cues to suggest malevolence. The other third
were also of a malevolent nature and low authority (i.e., contained no authority details
relating to the source of the pop-up such as company name, logo, or website link).
Participants had to decide whether to accept or decline pop-ups, at which point the
primary task would be reinstated at the point of interruption. Predictions informed by
parameters of SCAM [3] were supported, with an alarming 63% of mimicked pop-ups
accepted compared with 66% in the genuine condition. Even more worrying was that
56% of low authority pop-ups were accepted. Participants spent on average
only *5.5–6-s viewing pop-up message content before committing to a response.
When there were no time constraints to resume an interrupted task, participants
accepted a slightly higher percentage (72%) of genuine pop-ups and slightly fewer
(55%) mimicked pop-ups. This suggests that even without other cognitive and time
pressures, people are still not very good at detecting malevolent cues within mimicked
pop-up interruptions. [1] reported similar findings with older adults. Participants
demonstrated higher levels of susceptibility to malevolent pop-ups during an inter-
rupted memory recall phase, despite spending significant more time (*10.5–11-s)
viewing them than in [1]. Fitting with SCAM-based low suspicion and automaticity
predictions [3], both studies demonstrate very high levels of human susceptibility to
malevolent pop-up interruptions that occur during a demanding memory-based task.
However, concerns remain as neither study showed marked malevolent detection
improvements when time pressure was not a factor.
Given these results, it is important further develop and test interventions to reduce
susceptibility to computer-based communications such as malevolent pop-up messages.
Education-based training interventions are not always effective [20] with some finding
that people are more suspicious of scams that they are familiar with versus those that
are less familiar [21]. [22] tested the effectiveness of emails containing cues to
malevolence although found that not all people read and processed the content to a
deep enough level to identify them effectively. These findings fit SCAM parameters
regarding the use of automatic heuristic processing strategies, especially when
6 P. L. Morgan et al.
2 Method
2.1 Participants
Fifty Cardiff University Psychology undergraduate students (age: 19.32; SD 1.06) were
recruited, via opportunity sampling, in return for course credits with adequate a priori
power (.8 detect medium to large effect sizes (Cohen’s f .25 −.4). Participants were
first-language English or highly proficient in English as a second language and had
normal/correct vision. They were assigned to one of three cue identification training
groups. There were 16 in the Non-Malevolent Cue Identification (N-MCIT)/Control
group (M age: 19.63-years, four male), 17 in the Non-Incentivized Malevolent Cue
Identification (N-IMCIT) group (M age: 19.06-years, six male), and 17 in the Incen-
tivized Malevolent Cue Identification (IMCIT) group (M age: 19.29-years, two male).
2.2 Design
A mixed factorial design was employed. The between-participants’ independent vari-
able (IV) was CIT Group with three levels: Control, N-IMCIT, and IMCIT. There were
three repeated measures IVs. One was serial recall phase with two levels: Phase 1/Pre-
Intervention 1, and, Phase 3/Post-Intervention. Another was the malevolency (Message
Attempting to Reduce Susceptibility to Fraudulent Computer Pop-Ups 7
2.3 Materials
Prior to the start of Phase 1 and 3 trials, the following message was displayed in the
middle of the computer screen for 15-s:
‘This system is protected by virus protection software and pop-ups are installed on a regular
basis. However, please be vigilant about the security of this system by ensuring that any
attempts by applications to access system information of data are legitimate.’
2.4 Procedure
Before providing consent, participants read through an information sheet and experi-
mental instructions (which were also verbally read by the experimenter) before com-
pleting two practice trials: one with a non-interrupted serial recall task, and another
with a serial recall task interrupted by a non-malevolent pop-up. They were not
informed about the cyber security element of the experiment during this process. At the
beginning of Phase 1, participants were presented with the computer security message
(see Materials). After this disappeared, they were able to press the spacebar to start trial
one of 18, with 12 of the trials interrupted (see Materials). Phase 2 was the intervention
phase. Participants read an instruction sheet appropriate for their group. All were
instructed they had 5-min to read 5-passages (one-at-a-time) and complete the cue
Attempting to Reduce Susceptibility to Fraudulent Computer Pop-Ups 9
identification task relevant to their group. The Control group had to indicate (Yes or
No) whether the passage of text contained at least one cue relating to its category
description (e.g., color: look for color words). If answering yes, they then had to
indicate how many category words they could identify within the passage (i.e., 1–3). N-
IMCIT and IMCIT groups were first given written information pertaining to the
malevolency cues contained within pop-ups experienced in Phase 1. These were
explained verbally by the experimenter who checked participants’ understanding. As
with the Control group, participants in the MCIT groups were then presented with 5-
passages of text, one-at-a-time, and had to indicate (Yes or No) whether the passage it
contained at least one trained cue indicating potential malevolence. Participants were
also provided with a small whiteboard and marker to make notes, if desired. Phase 3
(post-intervention) involved 30 serial recall trials with 24 interrupted. After Phase 3,
participants completed demographics and pop-up awareness questionnaires. Partici-
pants were debriefed, with information about cyber-security and awareness aims.
All analyses are two-tailed with a = .05. One dataset was excluded, as it was found to
be a statistical outlier (z-scores > 3.29, ps < .001) on more than one measure.
Percentage of Pop-Up Messages Accepted/Declined
First, we consider mean percentages of ‘malevolent’ pop-ups accepted across Phases 1
(pre-intervention) and 3 (post-intervention), collapsing across New and Old cue
malevolent pop-ups in Phase 3 (Table 1). The percentage of malevolent pop-ups
accepted looks to have decreased in Phase 3 for both MCIT groups, although increased
for the Control group. Somewhat surprisingly, the mean percentage is markedly lower
in the Control versus the N-IMCIT and IMCIT groups.
Table 1. Percentage of Malevolent and Genuine pop-ups accepted during Phases 1 and 2 and
across each Training Group. Note. SD = Standard Deviation.
Malevolent Pop-Ups Genuine Pop-Ups
Phase Condition Mean SD Mean SD
1 Control 56.30 .34 63.54 .39
N-IMCIT 73.41 .31 84.31 .30
IMCIT 81.29 .29 87.25 .29
3 Control 67.69 .33 70.83 .35
N-IMCIT 60.35 .31 85.29 .24
IMCIT 70.12 .31 92.65 .11
however a significant interaction, F(2, 47) = 3.44, MSE = .04, p = .04. Bonferroni pot-
hoc tests revealed a non-significant (although trend) reduction in the percentage of
malevolent pop-ups accepted in Phase 3 compared with Phase 1 for the IMCIT group
(p = .07). However, the significant interaction might be better explained by the per-
centage of malevolent pop-ups accepted by the Control group in Phase 1 being sig-
nificantly lower than in the N-IMCIT and IMCIT groups within Phase 1 (ps < .025).
Given this unexpected difference (discussed later), another mixed ANOVA, this time 2
(Training Group: MCIT, IMCIT) 2 (Phase: 1, 3), was conducted. This revealed a
significant main effect of Phase, F(1, 32) = 5.63, MSE = .04, p = .02 with a lower
percentage of malevolent pop-ups accepted in Phase 3 than in Phase 1. There was a non-
significant main effect of Training Group, F(1, 32) = .96, MSE = .08, p = .33, and a
non-significant interaction, F(1, 32) = .03, MSE = .04, p = .86.
Taken together, these findings suggest that: (1) MCIT worked in terms of reducing
the percentage of malevolent pop-up messages accepted post-intervention, (2) IMCIT
did not lead to better performance than N-IMCIT, and, (3) participants in the Control
group, in Phase 1 at least, performed differently (i.e., chose to accept far less malev-
olent pop-ups) to those in MCIT conditions. In relation to (1), findings are in line with
SCAM predictions that heightening suspicion will lead to increased cognitive and less
automatic processing of stimuli [3], thus improving the likelihood of identifying
malevolence cues. However, the percentage of malevolent pop-ups accepted was still
very high, even after the intervention. In relation to (2), incentivized MCIT through
social comparison (using an onscreen leaderboard technique), was not effective enough
to cause even more suspicion and increased cognitive processing of potential cues to
suggest malevolence within pop-up messages compared to non-incentivized MCIT.
This finding (despite there being a trend) is not in line with [22] and possible reasons
are considered in the Limitations section. Considering (3), the only difference was
when the groups were tested: The Control group were tested after the MCIT groups.
Next, we consider mean percentages of ‘genuine’ pop-ups accepted in Phases 1 and
3, noting again that both New and Old cue malevolent pop-up data are collapsed across
(Table 1). The percentage of genuine pop-ups accepted increased marginally in Phase 3
across all groups. However, and as with malevolent pop-ups, the mean percentage of
genuine pop-ups accepted in Phase 1 was markedly lower in the Control versus MCIT
groups. A mixed 3 2 analysis of variance (ANOVA) with Training Group as the
between-subjects variable and Phase revealed a marginally non-significant main effect
of Training Group, F(2, 47) = 3.12, MSE = .07, p = .054, and a non-significant main
effect of Phase, F(1, 47) = 2.57, MSE = .02, p = .12. There was a non-significant
interaction. However, these findings might again be affected by the unusual pattern of
data in the Control condition during Phase 1 compared to the MCIT condition.
Therefore, a 2 (Training Group: MCIT, IMCIT) 2 (Phase: 1, 3) mixed ANOVA was
conducted. There were non-significant main effects of Training Group, F(1, 32) < 1,
p = .50, and Phase, F(1, 32) < 1, p = .39, and a non-significant interaction, F(1,
32) < 1, p = .55.
Taken together, these findings suggest that (1) the ability to identify genuine pop-
up messages was high, (2) MCIT did not have any effect on this, and (3) participants in
the Control group, in Phase 1 at least, performed quite differently (i.e., accepted fewer
Attempting to Reduce Susceptibility to Fraudulent Computer Pop-Ups 11
Table 2. Percentage of Old and New pop-ups accepted during Phase 3 across each Training
Group. Note. SD = Standard Deviation.
Malevolent
Pop-Ups
Phase Condition Mean SD
Old Control 67.75% .34
N-IMCIT 58.82% .35
IMCIT 66.65% .34
New Control 67.75% .36
N-IMCIT 61.79% .31
IMCIT 73.35% .30
We anticipated that participants in both MCIT groups, and in particular the I-MCIT
group would be less likely to spot new cues. However, there is no statistical evidence to
suggest that any form of MCIT led to participants accepting more New messages,
despite an *11.5% higher acceptance of these in the IMCIT versus the N-IMCIT
condition in Phase 3. Of course, this could be a power issue, and future studies should
12 P. L. Morgan et al.
consider this before ruling out the possibility that MCIT will not put people at a dis-
advantage in terms of spotting malevolent cues that they have not be trained to identify,
Time to Accept/Decline Pop-Up Messages
Next, we consider the time taken at make an accept/decline response. Noting that the
time to accept/decline malevolent pop-ups was 5.37-s for younger adults in the [2]
study, and 10-92-s for older adults in the [1] study. In the same studies, the times to
accept genuine pop-ups were 5.47-s and 10.45-s respectively. Mean pop-up
accept/decline times for the current study are displayed in Table 3 (with one outlier
removed: z-scores > 3.29, p < .001). Malevolent and genuine pop-ups, accept/decline
times are noticeably lower (*1–2-s) than in e.g., [2]. Also, response times appear to
reduce for each Group in Phase 3 versus Phase 1. The third, and somewhat counter-
intuitive observation, is that response times are noticeably lowest (and very short) for
the Control Group (M 3.39 Phase 1, M 2.97 Phase 3).
Table 3. Time (seconds) before making an accept/decline response to Malevolent and Genuine
pop-ups during Phases 1 and 2 and across each Training Group. Note. SD = Standard Deviation.
Malevolent Genuine
Pop-Ups Pop-Ups
Phase Condition Mean SD Mean SD
1 Control 3.45 2.65 3.33 2.23
N-IMCIT 4.41 2.49 4.26 2.15
IMCIT 4.77 2.99 4.57 3.05
3 Control 3.08 1.82 2.85 1.87
N-IMCIT 3.66 1.57 3.51 1.55
IMCIT 4.54 3.07 4.05 2.31
and identify cues than in the MCIT groups. This was not the case. Also, their accep-
tance rate for malevolent pop-ups in Phase 3 increased by over 10% and the time taken
to accept/decline messages reduced by almost half a second. Upon closer inspection of
the data, three Control group participants almost always declined malevolent messages
compared with the others whose performance was largely in line with those in the
MCIT groups. However, they were not statistical outliers the p < .001 (z-
scores > 3.29) level.
4 Limitations
There are limitations. First, there was no statistical evidence to suggest that those in the
IMCIT group were better at identifying malevolent pop-ups than those in the N-IMCIT
group, despite a trend. Perhaps using a leaderboard with individual position increasing
after each task (e.g., 19th/20 after the first task, 1st after the last task) was not effective
enough. This may be influenced by some participants potentially being aware that they
were performing optimally and met with incongruent feedback to suggest otherwise.
Competing with other people in situ may have promoted stronger social comparison
and led to more intense cognitive processing strategies [3]. Second, within both MCIT
conditions, participants had to identify whether they detected malevolent cues and then
type a number corresponding to how many. This method meant that accuracy of
malevolent cue identical could not be measured. Third, participants had one-minute per
training task, only five tasks to complete, with each passage containing only three
malevolent cues. They were also aware that there would be a maximum of three
malevolent cues. This may not have been cognitively engaging enough. Finally,
Control group participants were treating pop-ups with higher levels of suspicion in
Phase 1. Ideally, this condition would be re-run to check for a possibly anomalous
effect.
5 Implications
References
1. Morgan, P.L., Williams, E.J., Zook, N.A., Christopher, G.: Exploring older adult
susceptibility to fraudulent computer pop-up interruptions. In: International Conference on
Applied Human Factors and Ergonomics, pp. 56–68. Springer, Cham (2018)
2. Williams, E.J., Morgan, P.L., Joinson, A.N.: Press accept to update now: individual
differences in susceptibility to malevolent interruptions. Decis. Support Syst. 96, 119–129
(2017)
3. Vishwanath, A., Harrison, B., Ng, Y.J.: Suspicion, cognition, and automaticity model of
phishing susceptibility. Commun. Res. 45(8), 1–21 (2016)
4. Anti-Phishing Working Group (APWG). https://www.antiphishing.org/resources/apwg-
reports/
5. Department for Culture, Media & Sport.: Cyber security breaches survey 2017. https://www.
gov.uk/government/uploads/system/uploads/attachment_data/file/609186/Cyber_Security_
Breaches_Survey_2017_main_report_PUBLIC.pdf
6. National Cyber Security Centre.: Weekly threat report, 30 June 2017. https://www.ncsc.gov.
uk/report/weekly-threat-report-30th-june2017
7. Perera, D.: Researcher: Sony hackers used fake emails, Politico. https://www.politico.com/
story/2015/04/sony-hackers-fake-emails-117200
8. Forbes Cyber Security report. https://www.forbes.com/sites/ellistalton/2018/04/23/the-u-s-
governments-lack-of-cybersecurity-expertise-threatens-our-infrastructure/#20d248be49e0
9. HM Government. National cyber security strategy 2016–2021. https://www.gov.uk/
government/uploads/system/uploads/attachment_data/file/567242/national_cyber_security_
strategy_2016.pdf
10. Conteh, N.Y., Schmick, P.J.: Cybersecurity: risks, vulnerabilities and countermeasures to
prevent social engineering attacks. Int. J. Adv. Comput. Res. 6(23), 31 (2016)
11. Downing, D., Covington, M., Covington, M., Barrett, C.A., Covington, S.: Dictionary of
Computer and Internet Terms. Barron’s Educational Series, New York (2000)
12. Daintith, J., Wright, E.: A Dictionary of Computing. Oxford University Press, Oxford (2008)
13. Norton How To 2018. https://us.norton.com/internetsecurity-how-to-the-importance-of-
general-software-updates-and-patches.html
14. Altmann, E.M., Trafton, J.G., Hambrick, D.Z.: Momentary interruptions can derail the train
of thought. J. Exp. Psychol. Gen. 143(1), 215–226 (2014)
15. Hodgetts, H.M., Jones, D.M.: Interruption of the Tower of London task: support for a goal-
activation approach. J. Exp. Psychol. Gen. 135(1), 103–115 (2006)
16. Monk, C.A., Trafton, J.G., Boehm-Davis, D.A.: The effect of interruption duration and
demand on resuming suspended goals. J. Exp. Psychol. Appl. 14(4), 299–313 (2008)
17. Altmann, E.M., Trafton, J.G.: Memory for goals: an activation-based model. Cogn. Sci. 26,
39–83 (2002)
18. Altmann, E.M., Trafton, J.G.: Timecourse of recovery from task interruption: data and a
model. Psychon. Bull. Rev. 14(6), 1079–1084 (2017)
19. Cacioppo, J.T., Petty, R.E., Feng Kao, C.: The efficient assessment of need for cognition.
J. Pers. Assess. 48(3), 306–307 (1984)
20. Anandpara, V., Dingman, A., Jakobsson, M., Liu, D., Roinestad, H.: Phishing IQ tests
measure fear, not ability. In: International Conference on Financial Cryptography and Data
Security, pp. 362–366. Springer, Berlin (2007)
21. Downs, J.S., Holbrook, M.B., Cranor, L.F.: Decision strategies and susceptibility to
phishing. In: Proceedings of the Second Symposium on Usable Privacy and Security,
pp. 79–90. ACM (2006)
Attempting to Reduce Susceptibility to Fraudulent Computer Pop-Ups 15
22. Kumaraguru, P., Sheng, S., Acquisti, A., Cranor, L.F., Hong, J.: Teaching Johnny not to fall
for phish. ACM Trans. Internet Technol. (TOIT). 10(2), 1–30 (2010)
23. Clifford, M.M.: Effects of competition as a motivational technique in the classroom. Am.
Educ. Res. J. 9(1), 123–137 (1972)
24. Aleman, J.L.F., de Gea, J.M.C., Mondéjar, J.J.R.: Effects of competitive computer-assisted
learning versus conventional teaching methods on the acquisition and retention of
knowledge in medical surgical nursing students. Nurse Educ. Today 31(8), 866–871 (2011)
25. Festinger, L.: A theory of social comparison processes. Hum. Relat. 7(2), 117–140 (1954)
26. Peirce, J.W.: PsychoPy—psychophysics software in Python. J. Neurosci. Methods 162(2),
8–13 (2007)
Cyber Resilient Behavior: Integrating Human
Behavioral Models and Resilience Engineering
Capabilities into Cyber Security
Abstract. Cybercrime is on the rise. With the ongoing digitization of our society,
it is expected that, sooner or later, all organizations have to deal with cyberattacks;
hence organizations need to be more cyber resilient. This paper presents a novel
framework of cyber resilience, integrating models from resilience engineering and
human behavior. Based on a pilot study with nearly 60 small and medium-sized
enterprises (SMEs) in the Netherlands, this paper shows that the proposed
framework holds the promise for better development of human aspects of cyber
resilience within organizations. The framework provides organizations with diag-
nostic capability into how to better prepare for emerging cyber threats, while
assuring the viability of human aspects of cyber security critical to their business
continuity. Moreover, knowing the sources of behavior that predict cyber resiliency
may help in the development of successful behavioral intervention programs.
1 Introduction
Every day, companies all over the world suffer cyberattacks. The number of cyberat-
tacks on organizations worldwide has been growing over the last six years to an
average of 2.5 attacks every week for large organizations [1]. Based on 2,182 inter-
views from 254 companies in seven countries, the Ponemon institute [1] calculated that
the average cost of cybercrime in 2017 was 11.7 million US dollars per organization.
These costs are internal, dealing with cybercrime and lost business opportunities, and
external, including the loss of information assets, business disruption, equipment
damage and revenue loss. With the ongoing digitization of our society, it is expected
that the number of cyberattacks and, consequently, the annual costs of cybercrime, will
increase rapidly over the next years.
1
https://www.itgovernance.co.uk/cyber-resilience.
18 R. van der Kleij and R. Leukfeldt
2 Resilience
Research on resilience has increased substantially over the past decades, following
dissatisfaction with traditional models of risk and vulnerabilities, which focus on the
results of adverse events [21]. Whereas traditional risk models have historically been
useful for many applications in the past, it is difficult to apply them to cyber risks [4,
22]. Traditional risk assessment approaches tend to break down when it is difficult to
clearly identify the threats, assess vulnerabilities, and quantify consequences [23, 24].
Cyber threats cannot be clearly identified and quantified through historical measures
due to the rapidly changing threat environment [4]. Resilience, however, focuses on the
ability to succeed under varying conditions.
Resilience was first used in medical and material sciences, relating to the ability to
recover from stress or strain [25]. More recently a wider concept of resilience has
emerged. Within the domain of health sciences, the Resilience and Healthy Ageing
Network defined resilience as “the process of negotiating, managing and adapting to
significant sources of stress or trauma. Assets and resources within the individual, their
life and their environment facilitate this capacity for adaptation and ‘bouncing back’ in
the face of adversity. Across the life course, the experience of resilience will vary”
([21], p. 2).
Matzenberger [25] defines resilience in a learning environment as the capacity or
ability of a system to persist after disturbance and to reorganize or emerge while
sustaining essentially the same function. In hazards research, resilience is understood as
the ability to survive and cope with a disaster with minimum impact and damage. It
holds the capacity to reduce or avoid losses, contain effects of disasters, and recover
with minimal disruptions [25, 26]. In resilience engineering, a new approach to risk
management, resilience is described at the generic system level as the intrinsic ability
of a system to adjust its functioning prior to, during, or following changes and dis-
turbances, so that it can sustain required operations under both expected and unex-
pected conditions [27]. Van der Beek and Schraagen [28] have defined resilience at the
team level in crisis management situations by expanding the ability of a system with
more relation-oriented abilities of leadership and cooperation.
The key term to all these definitions is the system’s ability to adjust its functioning.
Resilience represents the capacity (of an organizational system) to anticipate and
manage risk effectively, through appropriate adaptation of its actions, systems and
processes, so as to ensure that its core functions are carried out in a stable and effective
Cyber Resilient Behavior: Integrating Human Behavioral Models 19
relationship with the environment [29]. Although these definitions at a glance mainly
seem to differ in level of analysis, ranging from the individual via the team level to the
more generic organizational (system) level, Woods [5] argues that resilience is used in
four different ways: (1) resilience as rebound from trauma and return to equilibrium;
(2) resilience as a synonym for robustness; (3) resilience as the opposite of brittleness,
i.e., as graceful extensibility when surprise challenges boundaries; (4) resilience as
network architectures that can sustain the ability to adapt to future surprises as con-
ditions evolve. The implication of this partition is that one needs to be explicit about
which of the four senses of resilience is meant when studying or modeling adaptive
capacities (or to expand on the four anchor concepts as new results emerge) [5]. Not all
these uses of the label ‘resilience’ are correct according to Woods. Resilience as the
increased ability to absorb perturbations confounds the labels robustness and resilience.
Some of the earliest explorations of resilience confounded these two labels, and this
confound continues to add noise to work on resilience [5].
The broad working definitions of resilient performance can be made more precise
and operational by considering what makes resilient performance possible. Since
resilient performance is possible for most, if not all, systems, the explanation must refer
to something that is independent of any specific domain. Hollnagel [30] has proposed
the following four abilities necessary for resilient performance (see also [27]):
• The ability to Anticipate. Knowing what to expect or being able to anticipate
developments further into the future, such as potential disruptions, novel demands
or constraints, new opportunities, or changing operating conditions. This is the
ability to create foresight and to address the potential.
• The ability to Monitor. Knowing what to look for or being able to monitor that
which is or could seriously affect the system’s performance in the near term, pos-
itively or negatively. The monitoring must cover the system’s own performance as
well as what happens in the environment. This is the ability to address the critical.
• The ability to Respond. Knowing what to do, or being able to respond to regular and
irregular changes, disturbances, and opportunities by activating prepared actions or
by adjusting current modes of functioning. It includes assessing the situation,
knowing what to respond to, finding or deciding what to do, and when to do it. This
is the ability to address the actual.
• The ability to Learn. Knowing what has happened, or being able to learn from
experience, in particular to learn the right lessons from the right experience, suc-
cesses as well as failures. This is the ability to address the factual. Although this
capacity is often overlooked, it is a critical aspect of resilience. By learning how to
be more adaptable, systems are better equipped to respond when faced with some
sort of disruption [25].
Hence, the resilience of a system is defined by the abilities to respond to the actual,
to monitor the critical, to anticipate the potential, and to learn from the factual [27]. The
abovementioned abilities can be thought of together as a framework for identification
and classification of indicators [25]. The engineering of resilience comprises the ways
in which these four capabilities can be established and managed [27]. This is of
importance to organizations because being resilient can provide them with a compet-
itive advantage [31]. Resilient organizations may also contribute to a cyber-resilient
20 R. van der Kleij and R. Leukfeldt
community or to more cyber resiliency at the nationwide level. McManus [32] argues
that resilient organizations contribute directly to the speed and success of community
resilience. Without critical services provided by resilient organizations, such as
transport, healthcare and electricity, communities (or states alike) cannot respond or
recover (see also [33]).
Although these four capabilities have been under debate for quite some time, and
research at the employee [34] and team level [28] only partially support the four-
dimensional nature of the construct as proposed by Hollnagel [27], we feel that these
four dimensions have high face validity in the domain of cyber security. For instance,
the recently completed NIST framework for improving Critical Infrastructure Cyber-
security encompasses five similar functions [35]. These five functions are Identify,
Protect, Detect, Respond, and Recover. They aid an organization in expressing its
management of cybersecurity risk by organizing information, enabling risk manage-
ment decisions, addressing threats, and improving by learning from previous activities.
The functions also align with existing methodologies for incident management and help
show the impact of investments in cybersecurity. For example, investments in planning
and exercises support timely response and recovery actions, resulting in reduced impact
to the delivery of services.
We feel that the four abilities as proposed by Hollnagel [27] together seem to be
sufficient without being redundant. We see no need, for instance, to split the function
Anticipate into the NIST functions of Identify and Protect. We think that the Protect
function, in which appropriate safeguards are developed and implemented to ensure
delivery of critical infrastructure services, is a composite rather than a primary ability of
a resilient organization. Implementing appropriate safeguards is a combination of the
ability to Anticipate and to Learn, and possibly also the ability to Detect (see also,
[30]). Moreover, the Identify function has a strong focus on understanding the contexts
and the risks, while Anticipate also looks at opportunities and emerging threats. We
also feel that the Recover function, to develop and implement the appropriate activities
to maintain plans for resilience and to restore any capabilities or services that were
impaired due to a cybersecurity incident, is essential, but again, we do not think of this
ability as a primary function. This function is the consequence of another ability,
namely Learning. For sustainable recovery to take place within organizational systems,
organizations should have knowledge of what has happened, or should be able to learn
from experience.
3 Human Behavior
There is an important role for employees within organizations to help the organization
become more cyber resilient. In principle, one could easily envisage employees per-
forming Monitoring, Responding, Anticipating and Learning functions to maintain
resilience in cyber capabilities. To explain the four resilience functions from a
behavioral perspective, a comprehensive framework for understanding human behavior
can be applied that involves three essential conditions: Motivation, Opportunity, and
Ability (MOA) [36]. The cyber resiliency of an organization is in part determined by
employees’ motivation, opportunity, and ability to perform the four generic resilience
Cyber Resilient Behavior: Integrating Human Behavioral Models 21
5 Pilot Study
An initial version of the Cyber Resilient Behavior Questionnaire was refined based on
experts’ feedback and on validity testing of the survey on a small group of SMEs. The
revised instrument has 42 statements, measured on a six-point Likert-type scale with no
mid-point (ranging from strongly disagree to strongly agree). Even-numbered Likert
scales force the respondent to commit to a certain position even if the respondent may
not have a definite opinion [40]. Statements assess capacity, opportunity, and moti-
vation regarding the performing of resilient functions to anticipate, monitor, respond,
and learn within the organizational context. The statements focus predominantly on
positive protective intentions of employees. This focus was chosen because respon-
dents are probably willing to reveal these behaviors in a survey, yielding usable and
reliable data (see also [19]). To reduce the chance of response bias, which is the
tendency to favor one response over others [41], an - ‘I do not know’ - option was
included for each statement. To further avoid response bias, both positively- and
negatively-worded items were used in the survey [42]. Negatively-worded items may
act as “cognitive speed bumps that require respondents to engage in more controlled, as
opposed to automatic, cognitive processing” [43].
Cyber Resilient Behavior: Integrating Human Behavioral Models 23
Table 1. Conceptual framework of resilient behavior. The left column shows the four generic
resilience functions. The consecutive columns specify the abilities for resilient behavior for each
of the three sources of behavior
Capability Opportunity Motivation
Anticipate Knowing Having resources to look for Willing to look for potential
what to developments further into the disruptions, novel demands or
expect future constraints, new opportunities,
or changing operating
conditions
Monitor Knowing Having resources to monitor Willing to monitor that which
what to the system’s own performance is or could seriously affect the
look for as well as what happens in the system’s performance in the
environment near term, positively or
negatively
Respond Knowing Having resources that help in Willing to respond to regular
what to taking prepared actions and irregular changes,
do disturbances, and opportunities
Learn Knowing Having resources to learn the Willing to learn from
what has right lessons from the right experience
happened experience
Next, a pilot study was conducted, and the results were examined to identify any
remaining problematic items and to establish the reliability of the main components of
the survey. A total of 56 SME employees completed the pilot version of our instrument.
All were high-level representatives at different SMEs. Cronbach’s alpha was used as a
measure of the internal consistency of the survey. This refers to the degree to which the
items measure the same underlying construct, and a reliable scale should have a
Cronbach’s alpha coefficient above 0.70 [44]. Cronbach’s alpha coefficients for each of
the four main functions (i.e., Anticipate, Monitor, Respond, and Learn) all exceeded
this recommended value. A series of Pearson product moment correlations were cal-
culated to further assess the relationship between the items used to create the three main
constructs. An examination of the correlation matrices revealed that all items signifi-
cantly correlated at 0.3 or above with the other items in that construct.
Although the main focus of the pilot study was to test the instrument and the
framework, the survey also included a set of questions concerning the incidence of
cybercrime and victimization. Respondents were asked if their SME had been con-
fronted over the last 12 months with cyberattacks and whether harms or costs were
involved. We now present some preliminary results from our survey.
Almost half of the SMEs in our sample (48% or 22 SMEs) had been the victim of at
least one cyberattack in the last 12 months. Phishing (32%) and viruses (11%) were
reported most. Seven SMEs (12%) reported that damage was caused, for instance in the
form of financial damage or business disruption. The overall score for cyber resilience
of the SMEs was considered poor to average. SMEs scored 3.5 on the six-point Likert
type scale. SMEs in our sample were best at responding to cyberattacks, and worst at
learning from attacks (3.7 and 3.2, respectively). Because of the small sample size, no
24 R. van der Kleij and R. Leukfeldt
analyses were performed at the behavioral level for each of these functions. Never-
theless, this pilot study clearly shows that there is ample room for improvement in the
cyber resiliency of the SMEs in our sample.
The cyber security field is in need of techniques to evaluate and compare the security
design of organizations [8]. Many techniques have been proposed and explored, but
these typically focus on auditing systems and technologies rather than on people. Our
work is aimed at measuring cyber resilience of organizations through its employees
rather than just with the technologies on which they rely [33]. Our framework gives an
overview of relevant cyber resilient behaviors of employees within organizations.
Accordingly, our conceptual framework allows for better development of human
aspects of cyber resilience. It provides organizations with diagnostic capability to better
prepare themselves for emerging cyber threats, while assuring the viability of those
cyber assets critical to their business continuity [45]. Moreover, knowing what sources
of behavior play a role in being cyber resilient, may help in the development of
successful behavior change intervention programs. In future work, investigating how to
link behavior change interventions to resilient behavior of employees, might prove
important.
The Cyber Resilient Behavior Questionnaire is intended to investigate the capa-
bilities, opportunities and motivation of people from all levels and functions of the
organization. Many cyber security measurement tools used by organizations rely on
information from only one or few organizational members, often specialists or man-
agers responsible for cyber security [33]. This produces biased results, based on a
single or few experiences, often with a vested interest in the results or scores achieved.
The results that are produced with the Cyber Resilient Behavior Questionnaire are
based on responses from a significant number of the company’s employees. It is
therefore more likely to tell us what the organization is actually doing and whether
measures and policies have been embedded in the organization’s social system [33].
However, in our pilot study, for practical reasons, only responses from high-level
representatives from a small sample of SMEs were collected. Future research would
benefit from the use of a larger sample of employees from all levels within organi-
zations. Moreover, the SMEs in our sample were mostly small retailers. It is essential to
validate our framework and instrument within other categories of the economy as well,
for instance with large businesses in the industrial or financial sector.
Funding and Acknowledgments. This work was partially supported by the municipality of The
Hague. The authors would like to thank Dr. Susanne van ‘t Hoff - de Goede, Michelle Ancher,
Iris de Bruin and students from HBO ICT at THUAS for their assistance with this research effort.
Further we would like to thank Dr. Jan Maarten Schraagen and Dr. Heather Young for their
thoughtful and detailed comments that greatly improved the quality and readability of the
manuscript. We are also grateful to the SMEs who agreed to participate in the surveys.
Cyber Resilient Behavior: Integrating Human Behavioral Models 25
References
1. Ponemon Institute: Cost of cybercrime study (2017). https://www.accenture.com/
t20171006T095146Z__w__/us-en/_acnmedia/PDF-62/Accenture-2017CostCybercrime-US-
FINAL.pdf#zoom=50
2. Ponemon Institute: 2016 Cost of Cyber Crime Study & the Risk of Business Innovation
(2016). https://www.ponemon.org/local/upload/file/2016%20HPE%20CCC%20GLOBAL%
20REPORT%20FINAL%203.pdf
3. Accenture: Gaining ground on the cyber attacker. State of Cyber Resilience (2018). https://
www.accenture.com/t20180416T134038Z__w__/us-en/_acnmedia/PDF-76/Accenture-
2018-state-of-cyber-resilience.pdf#zoom=50
4. DiMase, D., Collier, Z.A., Heffner, K., Linkov, I.: Systems engineering framework for cyber
physical security and resilience. Environ. Syst. Decis. 35(2), 291–300 (2015)
5. Woods, D.D.: Four concepts for resilience and the implications for the future of resilience
engineering. Reliab. Eng. Syst. Saf. 141, 5–9 (2015)
6. Brown, C., Seville, E., Vargo, E.: Measuring the organizational resilience of critical
infrastructure providers: a New Zealand case study. Int. J. Crit. Infrastruct. Prot. 18, 37–49
(2017)
7. Parsons, K.M., Young, E., Butaviciu, M.A., Mc Cormac, A., Pattinson, M.R., Jerram, C.:
The influence of organizational information security culture on information security decision
making. J. Cogn. Eng. Decis. Mak. 9(2), 117–129 (2015)
8. Bowen, P., Hash, J., Wilson, M.: Information Security Handbook: A Guide for Managers-
Recommendations of the National Institute of Standards and Technology (2012)
9. Cain, A.A., Edwards, M.E., Still, J.D.: An exploratory study of cyber hygiene behaviors and
knowledge. J. Inf. Secur. Appl. 42, 36–45 (2018)
10. Yoon, C., Hwang, J.W., Kim, R.: Exploring factors that influence students’ behaviours in
information security. J. Inf. Syst. Educ. 23(4), 407 (2012)
11. Leukfeldt, E.R.: Phishing for suitable targets in the Netherlands: routine activity theory and
phishing victimization. Cyberpsychol. Behav. Soc. Netw. 17(8), 551–555 (2014)
12. Leukfeldt, E.R., Kleemans, E.R., Stol, W.P.: A typology of cybercriminal networks: from
low-tech all-rounders to high-tech specialists. Crime Law Soc. Change 67(1), 21–37 (2017)
13. Young, H., van Vliet, T., van de Ven, J., Jol, S., Broekman, C.: Understanding human
factors in cyber security as a dynamic system. In: International Conference on Applied
Human Factors and Ergonomics, pp. 244–254. Springer, Cham (2018).
14. Bulgurcu, B., Cavusoglu, H., Benbasat, I.: Roles of information security awareness and
perceived fairness in information security policy compliance. In: Proceedings of the AMCIS,
pp. 419–430 (2009)
15. Dodge, R.C., Carver, C., Ferguson, A.J.: Phishing for user security awareness. Comput
Secur. 26(1), 73–80 (2007)
16. Talib, S., Clarke, N.L., Furnell, S.M.: An analysis of information security awareness within
home and work environments. In: Proceedings of the International Conference on
Availability, Reliability, and Security, pp. 196–203 (2010)
17. Crossler, R.E., Bélanger, F., Ormond, D.: The quest for complete security: an empirical
analysis of users’ multi-layered protection from security threats. Inf. Syst. Front. 1–15 (2017)
18. Da Veiga, A., Eloff, J.H.: A framework and assessment instrument for information security
culture. Comput. Secur. 29(2), 196–207 (2010)
19. Stanton, J.M., Stam, K.R., Mastrangelo, P., Jolton, J.: Analysis of end user security
behaviors. Comput. Secur. 24(2), 124–133 (2005)
26 R. van der Kleij and R. Leukfeldt
20. Winnefeld Jr., J.A., Kirchhoff, C., Upton, D.M.: Cybersecurity’s human factor: lessons from
the Pentagon. Harv. Bus. Rev. 93(9), 87–95 (2015)
21. Windle, G., Bennett, K.M., Noyes, J.: A methodological review of resilience measurement
scales. Health Qual. Life Outcomes 9(1), 8 (2011)
22. Linkov, I., Anklam, E., Collier, Z.A., DiMase, D., Renn, O.: Risk-based standards:
integrating top–down and bottom–up approaches. Environ. Syst. Decis. 34(1), 134–137
(2014)
23. Cox Jr., L.A.: Some limitations of ‘‘risk=threat x vulnerability x consequence’’ for risk
analysis of terrorist attacks. Risk Anal. 28, 1749–1761 (2008)
24. Frick, D.E.: The fallacy of quantifying risk. Def. AT&L 228, 18–21 (2012)
25. Matzenberger, J.: A novel approach to exploring the concept of resilience and principal
drivers in a learning environment. Multicultural Educ. Technol. J. 7(2/3), 192–206 (2013)
26. Cutter, S.L., et al.: A place-based model for understanding community resilience to natural
disasters. Glob. Environ. Change 18(4), 598–606 (2008)
27. Hollnagel, E.: RAG – the resilience analysis grid. In: Hollnagel, E., Pariès, J., Woods, D.D.,
Wreathall, J. (eds.) Resilience Engineering in Practice. A Guidebook. Ashgate, Farnham
(2011)
28. Van der Beek, D., Schraagen, J.M.: ADAPTER: analysing and developing adaptability and
performance in teams to enhance resilience. Reliab. Eng. Syst. Saf. 141, 33–44 (2015)
29. McDonald, N.: Organisational resilience and industrial risk. In: Hollnagel, E., Woods, D.D.,
Leveson, (eds.) Resilience Engineering, pp. 155–180. CRC Press, Boca Raton (2006)
30. Hollnagel, E.: Introduction to the Resilience Analysis Grid (RAG) (2015). http://
erikhollnagel.com/onewebmedia/RAG%20Outline%20V2.pdf
31. Parsons, D.: National Organisational Resilience Framework Workshop: The Outcomes.
National Organisational Resilience Framework Workshop (2007). http://www.tisn.gov.au/
Documents/FINAL1Workshop.pdf. Accessed 22 Nov 2012
32. McManus, S., Seville, E., Vargo, J., Brunsdon, D.: Facilitated process for improving
organizational resilience. Nat. Hazards Rev. 9(2), 81–90 (2008)
33. Lee, A.V., Vargo, J., Seville, E.: Developing a tool to measure and compare organizations’
resilience. Nat. Hazards Rev. 14(1), 29–41 (2013)
34. Ferreira, P., Clarke, T., Wilson, J.R., et al.: Resilience in rail engineering work. In:
Hollnagel, E., Paries, J., Woods, D.D., Wreathall, J. (eds.) Resilience in Practice, pp. 145–
156. Ashgate, Aldershot (2011)
35. NIST: Framework for Improving Critical Infrastructure Cybersecurity, v 1.1, April 2018.
https://nvlpubs.nist.gov/nistpubs/CSWP/NIST.CSWP.04162018.pdf
36. MacInnis, D.J., Moorman, C., Jaworski, B.J.: Enhancing and measuring consumers’
motivation, opportunity, and ability to process brand information from ads. J. Mark. 55, 32–
53 (1991)
37. Michie, S., Van Stralen, M.M., West, R.: The behaviour change wheel: a new method for
characterising and designing behaviour change interventions. Implement. Sci. 6(1), 42
(2011)
38. Parsons, K., McCormac, A., Butavicius, M., Pattinson, M., Jerram, C.: Determining
employee awareness using the human aspects of information security questionnaire (HAIS-
Q). Comput. Secur. 42, 165–176 (2014)
39. Parsons, K., Calic, D., Pattinson, M., Butavicius, M., McCormac, A., Zwaans, T.: The
human aspects of information security questionnaire (HAIS-Q): two further validation
studies. Comput. Secur. 66, 40–51 (2017)
40. Brown, J.D.: What issues affect likert- scale questionnaire formats? JALT Test. Eval. SIG 4,
27–30 (2000)
Cyber Resilient Behavior: Integrating Human Behavioral Models 27
41. Randall, D.M., Fernandes, M.F.: The social desirability response bias in ethics research.
J. Bus. Ethics 10(11), 805–817 (1991)
42. Spector, P.E.: Summated Rating Scale Construction: An Introduction, no. 82. Sage,
Thousand Oaks (1992)
43. Chen, Y.H., Rendina-Gobioff, G., Dedrick, R.F.: Detecting Effects of Positively and
Negatively Worded Items on a Self-Concept Scale for Third and Sixth Grade Elementary
Students (2007). Online Submission
44. Cronbach, L.J.: Coefficient alpha and the internal structure of tests. Psychometrika 16(3),
297–334 (1951)
45. Linkov, I., Eisenberg, D.A., Plourde, K., Seager, T.P., Allen, J., Kott, A.: Resilience metrics
for cyber systems. Environ. Syst. Decis. 33(4), 471–476 (2013)
An International Extension of Sweeney’s
Data Privacy Research
1 Introduction
It is increasingly clear that two phenomena have grown extensively over the past two
decades: first, the exponential growth of cyber-attacks in virtually every computing
environment; and second, public awareness (whether accurate or not) of one’s vul-
nerability to attacks that may be directly aimed at the individual, or more generally to
an organization that maintains widespread data on the entire population.
The pioneering research of Dr. Latanya Sweeney demonstrated the vulnerability of
most residents and computer users in the United States to the easily available demo-
graphic data necessary to identify sensitive information about any individual [1]:
“It was found that 87% (216 million of 248 million) of the population of the United
States had reported characteristics that likely made them unique based only on {5-digit
ZIP, gender, date of birth}.” [1].
However conclusive was Sweeney’s research concerning the citizens and residents
of the United States, her research only provided a template for developing similar
estimates regarding other countries throughout the world.
It is our purpose to extend the previous research to develop similar estimates
regarding residents’ vulnerability to data attacks using similar demographic data. In the
first part, we will explore, by estimates of computer and Internet usage, and the
prevalence of cyber-attacks, in many world countries.
Based on such estimates of usage and vulnerability, we will explore comparable
demographic data for a number of target countries by utilizing population, life
expectancy, and postal coding systems in the selected countries.
The value of the Sweeney research has been to introduce residents of the United
States of the ease by which they can be identified in various databases and hence how
their personal information and be captured via techniques known generally as “social
engineering” or “dumpster diving.” Since approximately 87% of the US population can
be identified uniquely by only three pieces of (usually) easily found data, the identity of
the individual can easily be compromised by persons seeking that information in
publicly available databases.
In order to achieve the objectives of this paper, we will examine the feasibility of
obtaining similar data or persons in a selection of other countries.
We will begin by developing a methodology to select other countries for which this
analysis can be performed. First, consider the level of concern in countries throughout
the world in terms of the susceptibility to cyberattacks to discover personal informa-
tion. Although the level of attacks is rising in virtually every country, we postulate that
the level of concern by an individual citizen in a given country may be related to the
widespread availability of national computer usage and Internet usage. These data will
be demonstrated in Table 1 using data for the 30 countries with the greatest prevalence
of computer availability and Internet usage, both in terms of the total numbers and the
percentage of the population. It is part of our hypothesis that if a relatively small
percentage of a country’s population operate in cyberspace, there will be less interest
either among the country’s residents in terms of protecting their personal data; and by
the same token, interest amongst those involved in cyber-attacks in finding personal
data, since it might apply only to a very small percentage of the country’s population.
The methodology for the selection of countries to be considered in this analysis is
as follows: three statistics were identified for virtually all world countries. The first is
each country’s current population; the second, the percentage of the population owning
or using computers; and third, the percentage of the population with access to the
Internet. The use of these three statistics will generally give some measure as to the
possibility of identity theft, whether done locally or via the Internet. For each of these
three statistics, the countries of the world are ranked from the highest value to the
lowest; then, if for country X, Xpoprank will represent the country’s rank in world
population, from highest to lowest among world countries [2]; Xinternetuse, the per-
centage of Internet usage in the country, also from highest to lowest [3]; and finally
30 W. Patterson and C. E. Winston-Proctor
In order to determine a rank order for countries being considered, define a devel-
opment index based on the rank country among world countries using three orderings:
first, the rank of world countries by population (2018 projections); second, the per-
centage of the population with Internet access; and third, the percentage of the popu-
lation with local computer usage. To develop the overall ordering or countries that we
An International Extension of Sweeney’s Data Privacy Research 31
wish to analyze, each of the three orderings above is added for each country, then all
countries in our consideration are ranked by the lowest of the sums. In other words the
optimal rank for a country would be if it was ranked first in each of three categories,
and thus the sum of the three. In this analysis, the first country in combined ordering is
Japan, with its three indicators adding to 31.
Table 2. (continued)
12 Russia 2% 18 7 7 17 14
13 Canada 2% 5 40 3 14 10
14 South 2% 21 19 4 15 7
Korea
15 Taiwan 2% 11 21 12 11 15
16 Japan 2% 7 29 11 22 11
17 Mexico 2% 6 18 31 21 16
18 Argentina 1% 44 12 20 12 18
19 Australia 1% 14 37 17 27 13
20 Israel 1% 40 16 15 16 22
Using the data from the original 30 countries selected, as well as the additional 22
from the Symantec data, we will attempt to perform a simplified analysis, comparable
to the work by Sweeney, in order to develop national comparisons of the percentage or
numbers of persons in the population of each country in terms of the uniqueness of
identification, using factors similar to Sweeney’s: gender, date of birth (including year),
and postal code of residence.
It should be noted that in a number of the countries that we select in terms of the
criteria described above, there is no postal code system, and thus a comparable analysis
cannot be performed. We note that for our further analysis, we choose (for space
reasons) the leading 20 countries identified in Table 1 in terms of attack vulnerability.
In addition to these 20, we note that 18 also appear in Table 2, for the actual sus-
ceptibility to attack. However, it is noted that all of the leading 20 countries in Table 1,
the only countries in Table 2 but not in Table 1 are India and Israel.
To replicate the Sweeney study for other countries, it is necessary to identify the total
population, the life expectancy by country, and the postal code system in such
countries.
The first two are easily found and have a high degree of accuracy. The existence of
the postal code system, which does exist in most countries but not all; is of a different
nature, since the information that is easily available is the potential range of values for
postal codes in all of our selected countries. For example, and using ‘N’ to represent
decimal digits in a postal code, and ‘A’ for alphabetical characters, it is possible to
determine the total range of possible postal codes. For example, in the United States
five-digit ZIP code system, which we would indicate as “NNNNN”, there are a total of
105 = 100,000 possible postal codes. However, as reported by Sweeney at the time of
her research, only 29,000 of the possible five-digit combinations were actually in use.
(The corresponding use of US ZIP code numbers at present is 40,933.)
An International Extension of Sweeney’s Data Privacy Research 33
Most of these data have been compiled for approximately 200 countries, but in
order to apply the Sweeney criteria, we limit the further analysis to a smaller set of
countries.
In order to develop a comparison in terms of the privacy considerations in indi-
vidual countries, is necessary to be able to estimate the key statistics Sweeney used.
Population data is easily available for all United Nations member countries, as are
mortality rates or life expectancy rates to develop applicable birthdays as in Sweeney’s
paper. However, the third statistic used by Sweeney is not as widely available. This
statistic is, for the United States, the 5-digit form of postal code, called in the US the
“ZIP code”. It is noted that in the US context, that most if not all postal service users
have a 9-digit ZIP Code, sometimes called the “ZIP+4”, NNNNN-NNNN, but the
recording and storage of US postal codes still varies widely, and most databases that
might be discovered by a hacker would only have the 5-digit version, “NNNNN”.
Furthermore, Sweeney’s original research only used the original 5-digit ZIP Code.
In our comparative study, we have had to eliminate a number of United Nations
countries that either do not have the postal code system, or it is not readily available.
This seems to be only in a relatively small number of United Nations countries.
The other complicating factor in this comparative study is that in most countries,
there is a distinction between the characters of potential postal codes as a function of
the syntax of the structure of postal code assignment. Throughout the world, most
postal codes use a combination of numerals {0, 1, 2, …, 9} which we describe as ‘N’;
and letters of the relevant alphabet. In the Roman alphabet (mostly in uppercase), we
have {A, B, C, …, Z} which we designate as ‘A’.
In the case of the older US 5-digit ZIP Code, the syntax is NNNNN, which allows
for the maximum possible number of ZIP Codes as 105 = 100,000. As a comparison,
the Canadian postal code system is ANA NAN, therefore 263 103 = 17,576,000.
Thus our level of analysis in estimating the actual number of postal codes is simply
to use the calculated level based on the syntactical postal codes. To be more accurate,
however, it is necessary to take into account that many postal systems restrict the usage
of some of these symbols for perhaps local reasons. Thus, to obtain a more precise
comparison, it is important for possible to determine the actual number of postal code
values actually in use, as opposed to the number theoretically in use.
For example, the current estimate of US ZIP Code numbers in use is 40,933, or
41% of the allowable values. These estimates are only available for a smaller number of
countries.
In order to determine a “Sweeney Index” for our 32 selected countries (30 from
Table 1, India and Israel from Table 2), we must first determine the life expectancy by
country, and the number of possible values in the country’s postal code system.
The first challenge in this analysis arises because not all countries have a postal
code system. The following table demonstrates the status of postal codes in our 30
selected countries. In most cases where a postal code system exists, it is defined by a
numerical sequence, ranging from four, five, six or even nine digits; and often also by
several alphabetic symbols, the most part using the Roman alphabet. In the table below
the use of a numeric character is indicated by N, and an alphabetic character by A.
Thus, for example, a country using a five-digit number for the postal code would be
represented in our table as “NNNNN”.
34 W. Patterson and C. E. Winston-Proctor
The first estimate of the number of postal codes for country is determined by the
syntax and the potential number of occurrences for each character in the string rep-
resenting the code. In a number of cases, it is possible to determine if a country uses all
of the possibilities for codes under its coding system. But in most countries, not all
possibilities are used—only a certain fraction of the eligible set of codes are actually in
use; unfortunately this information is not readily available for all countries.
The major conclusions by Sweeney are obtained by the analysis of internal United
States data on ZIP Codes to approximate the distribution of active addresses as dis-
tributed over the entire set of postal code values. Sweeney defines several methods of
distribution, including uniform distribution, which would certainly simplify calcula-
tions for other countries. It is likely to be less realistic than many other options;
nevertheless, the scope of this article, we will only calculate the likelihood of unique
identification of individuals assuming uniform distribution of individuals in countries;
since we do not have access to the necessary internal postal code distributions in other
countries. Nevertheless, we feel uniform distribution gives a reasonable first approxi-
mation to the Sweeney results.
Using the uniform distribution process, we can calculate the total number of “pi-
geonholes” accurately for many countries, and then the uniform distribution by
dividing the population by the number of pigeonholes.
5 “Pigeonholes”
The problem then becomes the conducting of an assessment of the data for the number
of persons that can fit into each of the potential categories, or “pigeonholes” in a
frequently-used term in computer science. Another way of phrasing the conclusions of
Sweeney’s earlier study is to say that of all the pigeonholes, approximately 87% have
no more than one datum (that is, no more than one person) assigned to that pigeonhole.
Table 3. (continued)
Country Postal code Maximal possible Actual postal National
format postal codes codes (pp) population
Spain NNNNN 100,000 56,542 46,354,321
Brazil NNNNNNNN 100,000,000 5,525 209,288,278
Australia NNNN 10,000 2,872 24,450,561
Mexico NNNNN 100,000 100,000 129,163,276
China NNNNNN 1,000,000 860,000 1,409,517,397
Netherlands NNNN AA 250,000 5,314 17,035,938
Malaysia NNNNN 100,000 2,757 31,624,264
Argentina ANNNAAA 240,000 1,237 44,271,041
Philippines NNNN 10,000 10,000 104,918,090
Poland NN-NNN 100,000 21,965 38,170,712
Italy NNNNN 100,000 4,599 59,359,900
Turkey NNNNN 100,000 3,314 80,745,020
India NNNNNN 1,000,000 153,343 1,339,180,127
Israel NNNNNNN 10,000,000 3,191 8,321,570
The number of pigeonholes in Sweeney’s study for the United States is calculated
by the product of the potential number of persons identified by birthdate including year,
gender, and 5-digit ZIP code. The contribution to this number related to gender is 2, say
pg = 2. For birthdate, we approximate the number of values using 365 for days of the
year (a slight simplification ignoring leap years), multiplied by the number of years,
estimated by the country’s life expectancy in years [5]. Call this pb. The final relevant
factor in estimating the number of pigeonholes is the number of potential postal codes,
pp. Then the total number of pigeonholes is
One remaining problem is the calculation of the number of postal codes, pp [6]. It is
an easy calculation to find the maximal value for pp say ppmax. For example, for the 5-
digit US ZIP code system, that maximal value is ppmax = 105 = 100000. At the time of
Sweeney’s research, the number of ZIP codes actually used was pp = 29343 ([1], page
15), or 29.3% of the total number of ZIP code values. At present, the number of ZIP
codes in use is 40,933.
Given available data for all world countries, the value pp is often not made public
(Table 3).
It is not possible with most countries perform the same detailed analysis as can be
found in the work of Sweeney, since for many countries the detailed data she obtained
is not available. However, several other analyses are possible which can still provide
36 W. Patterson and C. E. Winston-Proctor
considerable insight with the comparable Sweeney analyses and hence the state of
privacy issues in certain countries.
The two analyses we have performed use the following:
1. Assumption of uniform distribution of Sweeney parameters with postal codes using
theoretical postal code values.
2. Assumption of uniform distribution of Sweeney parameters with postal codes using
actual postal code values.
We have been able to perform the analyses for most United Nations countries
where appropriate postal information is available. However, for the purposes of this
paper, we have restricted to those 30 countries we identified in Sect. 1 with the
potential of the greatest risk because of greatest computer and Internet penetration.
7 Conclusions
The statistic in the final column of Table 4 provides the key information for our
conclusions. The meaning of the statistic “Average Number of Persons/Pigeonhole”,
which we will abbreviate AvgPP, is an indication of the potential lack of individual
privacy for computer users in the country in question. We note first that the calculation
for the US, 0.1369, is the level by which a single individual may not be identified; thus
1 − AvgPP = 0.8631 is very close to the statistic computed by Sweeney in her 2001
paper, that 87% of the US population could be uniquely identified by the Sweeney
criteria; in other words, only 13% of the US population could not be so identified.
Rather than a direct confirmation, this does demonstrate a consistency even with an 18-
year difference in the date and a change in the US population (248,418,140 in 2000 and
324,459,463 at present) and in the number of used ZIP codes (29,343 in 2000 and
40,933 at present).
Table 4. Average number of residents per Sweeney criterion (pigeonholes) for 22 countries
Country Gender Life Actual Pigeonholes Population Average no. of
x expectancy postal persons/pigeonhole
days/yr codes
Great 730 81.2 1,700,000 1.008E+11 6.62E+07 0.0007
Britain
Canada 730 82.2 834,000 5.005E+10 3.66E+07 0.0007
South 730 82.3 63,000 3.785E+09 5.10E+07 0.0135
Korea
Spain 730 82.8 56,542 3.418E+09 4.64E+07 0.0136
Mexico 730 76.7 100,000 5.599E+09 1.29E+08 0.0231
China 730 76.1 860,000 4.778E+10 1.41E+09 0.0295
Poland 730 77.5 21,965 1.243E+09 3.82E+07 0.0307
(continued)
An International Extension of Sweeney’s Data Privacy Research 37
Table 4. (continued)
Country Gender Life Actual Pigeonholes Population Average no. of
x expectancy postal persons/pigeonhole
days/yr codes
Japan 730 83.7 64,586 3.946E+09 1.27E+08 0.0323
Israel 730 82.5 3,191 1.922E+08 8.32E+06 0.0433
France 730 82.4 20,413 1.228E+09 6.50E+07 0.0529
Netherlands 730 81.9 5,314 3.177E+08 1.70E+07 0.0536
Russia 730 70.5 43,538 2.241E+09 1.44E+08 0.0643
United 730 79.3 40,933 2.370E+09 3.24E+08 0.1369
States
Australia 730 82.8 2,872 1.736E+08 2.45E+07 0.1408
Germany 730 81.0 8,313 4.915E+08 8.21E+07 0.1671
India 730 68.3 153,343 7.646E+09 1.34E+09 0.1752
Malaysia 730 75.0 2,757 1.509E+08 3.16E+07 0.2095
Philippines 730 68.5 10,000 5.001E+08 1.05E+08 0.2098
Italy 730 82.7 4,599 2.776E+08 5.94E+07 0.2138
Turkey 730 75.8 3,314 1.834E+08 8.07E+07 0.4403
Argentina 730 76.3 1,237 6.890E+07 4.43E+07 0.6425
Brazil 730 75.0 5,525 3.025E+08 2.09E+08 0.6919
The meaning of the ordering in the AvgPP Column is to show that in which
countries personal privacy is even less than in the United States (again 87% identifiable
by the three Sweeney criteria). These are the 12 countries above the United States in the
Table 4, with the Great Britain having the least down to Israel; and the other 9 below
the United States having a greater level of personal privacy, from Australia to Brazil.
This paper has implications for future behavioral cyber security problem specifi-
cation, theoretical conceptualization, and methodological development.
References
1. Sweeney, L: Simple Demographics Often Identify People Uniquely. Carnegie Mellon
University, Data Privacy Working Paper 3, Pittsburgh (2000)
2. United Nations. https://population.un.org/wpp/Download/Standard/Population/
3. Wikipedia. https://en.wikipedia.org/wiki/List_of_countries_by_number_of_Internet_users9)
4. Symantec, Internet Security Threat Report. ISTR 2018 vol. 3. http://resource.elq.symantec.
com
5. World Health Organization. http://apps.who.int/gho/data/node.main.688?lang=en
6. Wikipedia. https://en.wikipedia.org/wiki/List_of_postal_codes
The Human Factor in Managing the Security
of Information
1 Introduction
granted access to information required for the performance of their duties, and this can
also result in malpractice (mainly due to recklessness).
The most serious breaches invariably exploit the human factor, which means that
special emphasis must be placed on educating workers, and that security training
programs must also include all manipulative techniques, the so-called social engi-
neering techniques. In large companies, too much stress is placed on the sophisticated
security of information systems, while a wide range of aspects that depend solely on
individuals’ good will and common sense is neglected.
There is evident a somewhat perplexing paradox in all of this - the authorities of the
organization predominantly believe that they are more susceptible to attacks from
without and not to the breaches from within the organization. Such a claim is made
despite the fact that there have been made repeated and confirmed observations to the
effect that threats caused by authorised network users – permanent personnel as well as
temporary staff or former employees having access to resources located within the
companies - are more likely and more dangerous in their repercussions.
A question needs to be raised as to how real this threat is. There is a lack of official
data on losses caused by insiders’ unauthorized actions, but abuses affect practically
every company, regardless of geographical location, industry, size and they typically
remain undetected [7]. According to surveys conducted among employees in the course
of research on, among others, malpractice, one in five employees declares that they are
aware of other employees robbing the employer. As the enterprise grows, the possi-
bility of abuse increases while the detection rate drops [5].
The aim of this article is to present the levels of awareness of information security
risks in such specific organizations as higher education institutions. The initial research
was carried out in 2008, followed by a comparative study performed in the same
entities in 2018 with a view to verifying whether any changes have occurred. The data
from the latest survey will be presented and those from 2008 will be referred to for
comparison. The research focused on public universities in Poland. The study was part
of the research project of the SYDYN group.
2 Awareness of Threats
Safety is inextricably linked to costs, and universities are aware of the need to pay
more for higher quality products. They require IT solution providers to run more and
more rigorous tests to detect the vulnerability of their products before releasing them on
the market, rather than utilize end users as beta-testers who will find errors on their own.
Given that information is the most valuable commodity of our time, its theft has virtually
become an everyday practice. The victims of acts of information theft, are, first of all,
naivety and ignorance of the employees in an organization while various techniques of
manipulating personnel are the most frequently deployed weapon. The application of
trickery in order to extort business information from authorized employees is one of the
means to obtain an organization’s resources. Others employ more advanced techniques
to hack information systems, and others, still, lead to intrusions into buildings and
premises of the organization. Aside from the typical threats, both external and internal,
whose causes can be attributed to failures, disruptions, and errors in information sys-
tems, the primary source of threats is the human factor. It is individuals who have the
biggest impact on the system. Through their actions they can cause both hardware and
software malfunctions, breach data sets and access other system resources. The type and
magnitude of a potential threat depends on the function that a given person performs in
the organization, his or her skills and, of course, whether he or she acts unintentionally
or deliberately. Therefore, threats caused by human activities can be divided into
unintentional and deliberate. Conducted research has demonstrated that the vast
majority of unintentional threats (similarly as in 2008) are completely random and are
usually the result of inattention, submission to manipulation, thoughtlessness, negli-
gence or incompetence of individuals in the design or operation of the system.
Methods of manipulating personnel include, among others, sending falsified e-
mails and impersonating other users whom the individual trusts by entering the per-
son’s address in the identification field and then including a request or order to send
information or open attachments which, when activated, may initiate a process enabling
remote access to the user’s workstation (browsing catalogue contents, changing and
deleting information, intercepting passwords or encryption keys). Unintentional threats
are numerous and hardly predictable. Unfortunately, these threats cannot be eliminated
by applying technical solutions. These problems have to be solved procedurally. The
risk of their occurrence can be reduced by appropriate selection of staff, reasonable
allocation of access rights, trainings, software and hardware security measures.
Quite different is the case of intentional and deliberate threats. A very significant
hazard is internal attacks typically related to improper use of the organization’s ICT
system by users. Whereas the share of internal attacks in all recorded events is very high,
the only effective protection against such events is a well-designed security policy.
Only a handful of universities (10%) have begun work on developing a training and
awareness-building programs - a fundamental element in establishing an effective
information security strategy. There is no doubt that it is difficult for employees to react
to abuses committed in the organization if they have not been adequately trained in
recognising such incidents and reporting their occurrence to the right personnel.
42 M. Wisniewska et al.
In most cases, an organization’s management confirms that the users’ lack of risk
awareness is the main obstacle to achieving the required level of security.
Investment in technical solutions cannot completely eliminate the human risk
factor. Even if senior management claims that information security is important, time,
attention, actions taken and resources allocated to improve security are not adequate.
A number of observations of participants and interviews with representatives of
universities were conducted. They revealed several categories of irregularities typical
for the majority of the surveyed universities. In many cases, the respondents revealed
that they suspect, rather than know, that in some aspects of information security policy
certain steps are being implemented to the benefit of their units.
The results of research conducted by the authors for nearly 15 years indicate the
following issues:
– management is reluctant to attach adequate importance to human capital and is
much more willing to invest in technical solutions,
– fewer than half of the study participants provide systematic safety and control
training to their staff.
The most frequent irregularities in the area of information protection are identical to
those reported in 2008. It should be noted that despite considerable technical progress in
the field of information security and the declared increase in awareness among system
users and management, there has followed no significant improvement in security.
Among the most common sources of risk, the following have been recognized.
1. Incorrectly positioned screens at customer service workstations – computer
screens are positioned in such a way that unauthorized persons are able to read the
displayed information (e.g. in the customer queue, persons can see the data of the
customer being served and photograph them).
2. Employees do not lock workstations with password when leaving the work-
station – this gives unauthorized persons easy access to information contained in
the systems (especially in customer service rooms).
3. Employees do not lock (do not physically secure) rooms when leaving – there is
easy access to information contained both in IT systems (with unsecured work-
stations) and any paper documents located in the room. Documents with confi-
dential information about the organization as well as information concerning the
organization’s stakeholders may be stolen.
4. Documents are stored in generally accessible rooms - cabinets with documents
(frequently unsecured) stand in an area accessible to everyone entering the room.
Sometimes such cabinets stand outside a room, e.g. in a corridor. In customer
service rooms these are frequently not cabinets per se, but shelves lined with
documents.
5. 5. Internal attacks - involving the use of e.g. social engineering. There is lack of
awareness among employees that no information to which they have access should
be given or disclosed to unauthorised persons. Employees, in particular depart-
ments that process information that is not accessible to everyone (e.g. financial,
The Human Factor in Managing the Security of Information 43
human resources, IT, science, teaching, etc.), do not check the identity and
authorization of those who ask for information.
6. Movement around the organization’s premises - employees who have been
made redundant and have ceased cooperation with the organization should not
have access to office space. A barrier in the form of security is not effective if there
is no proper flow of information within the organization. Security personnel and
other employees are not informed that specific individuals no longer work for the
organization and have no right of access to certain areas.
7. Disclosure of information concerning employees of the organization - this
mainly involves posting internal information concerning the organization and
employees in publicly accessible areas.
8. Authorization regarding access to information systems - lack of procedures for
organizing matters related to granting, modifying and removing of authorization to
information systems.
a. Employees are granted or modified authorization to the systems without
written confirmation. There is no evidence that granting privileges has been
authorized. Unfortunately, such situations usually involve the most senior
officials in organizations - they usually have urgent needs and believe that no
regulations apply to them.
b. Clearances in IT systems and network resources are not withdrawn in due
time - e.g. upon termination of cooperation.
c. Lack of uniform rules on the storage of information on clearance levels - lack of
a uniform system of authentication and authorization.
d. The fourth case - authorizations for third parties. It is frequent that out-
sourced personnel, service technicians, contract workers, etc. perform work for
the organization. However, this area is unregulated, because e.g. the systems
themselves do not offer an option of granting rights in a manner consistent with
the internal procedures. In the majority of cases, there exist the so called special
accounts with wide range of access rights which are used by service techni-
cians, administrators, programmers, etc. This cannot be allowed to happen.
Accounts must be individually created for everyone who is supposed to have
access to the system - with a separate identifier. Otherwise, in the event of an
incident it is impossible to know who did what and when.
9. Sharing personal logins – this is also a very frequent situation: an employee
discloses his or her own system login and password to a colleague at work.
10. Storing system passwords - it is still true that employees store passwords to
systems on post-it notes under the keyboard, desk drawers are full of slips of paper
with specified systems, logins and passwords to them. There were reports of
notebooks with entire password history in these drawers.
11. Open wireless networks - in most units (70% vs. 85% in 2008) there were
examples of open wireless networks operating without any security measures. In
many cases the coverage of these networks extended even beyond the university
campus gounds.
44 M. Wisniewska et al.
public is easily excited about viruses and incidents involving external intrusions, it is
the employees and authorized network users who can cause the most damage.
The increased emphasis on the application of the company leadership’s security
management model, which includes raising employee security awareness, should also
entail a more thorough scrutiny of employees and business partners.
There is some irony to it. Human resources represent the greatest potential risk, but
when provided with appropriate support, they can become the strongest element of
protection within an organization. Assigning information security the right priority “at
the top” is a key element of success. The right approach and motivation at the highest
level of management can inspire radical changes. It is widely accepted that the per-
ception of information security as a necessary cost to business appears obvious.
However, looking at information security as a business growth factor, as an element of
competitive advantage, and as a method of protecting the value of resources, requires a
completely new and broader perspective. Senior management’s perception of infor-
mation security must therefore undergo a radical change. It is for university managers
to create a culture of security based on each unit’s awareness and sense of responsi-
bility for their own actions. Creating such an organizational culture is challenging and
can be achieved if senior management itself follows the rules it has established. If
organizations do not make information security a part of their culture, it is likely that
regulators will become involved in this process in the near future. With the increasing
complexity of the business world, information security requirements are also
increasing.
5 Conclusion
organization to function. The most effective tool is the Information Security Policy,
which may also include many other policies [8]. One such policy is, of course, the IT
Systems Security Policy. In addition to the management’s declaration of what it wants
to protect and for what reasons, this document also contains procedures, rules and
instructions, i.e. documents which should state how to fulfil the declarations.
Equally important as the implementation of appropriate technological solutions, the
creation and implementation of required procedures, such as training and raising
employees’ awareness of their responsibility for the security of processed information.
The starting point for effective information security management in the company is the
understanding that it does not consist only in the purchase and implementation of such
or other security systems.
The indispensability related to the implementation of the Information Security
Policy has its source in legal and business aspects, however, it is associated with the
need to protect the organization against various internal and external threats. The fact
that the threats are an inherent element of every area of the organization’s operation
should prompt the university authorities to reflect on an enormously important issue
related to their identification and identification of their sources. Most often, however, it
is so that the organization’s activities related to conducting a thorough threat analysis
are not the starting point for the creation of a security system. As a consequence, the
developed protection system is imperfect, not to say flawed, and, in any case, may
prove to be ineffective. This is due to the fact that if the most important weaknesses and
threats are not adequately known, the only outcome of the activities related to the
creation of a security system is to take into account the budget and qualifications of the
system’s developers, while neglecting the actual needs that result from the threats and,
as a result, affect the final form and effectiveness of the designed security system.
The IT risk is growing at an alarming rate as the authorization levels and the
number of users having access to information increase. Individuals operating within an
organization believe that the risk of being caught for fraud is low, as they have detailed
knowledge of the system, possible means of access and know organizational control
mechanisms. They usually work for personal gain, but some can also launch com-
pletely selfless attacks. The recklessness or mere ignorance of others, weakening even
the best controls and technical security mechanisms, may mean that such individuals
may become the greatest threat to information security within an organization.
Unfortunately, business managers too often underestimate this threat. They only
mention it as the second most important.
The lack of progress in information security in universities compared to 2008 levels
may be surprising. After all, more and more members of the general public use
advanced IT tools on their own and know how important it is to protect their identity,
image and financial assets, among others. Individuals know how to protect themselves
and they do it better and better. Why is there no significant progress in information
protection at universities? The probable reason is that in systems that we use privately
and for our own purposes, the technological advancement is so great that system
providers compete with each other very strongly as they are exposed to constant attacks
from the press and Internet users on social networking sites. Therefore, they try to
implement security mechanisms that are resistant to not only attacks, but also even to
accidental loss of security. Such actions have made us, as users of IT systems, less
The Human Factor in Managing the Security of Information 47
sensitive to threats. And this is paradoxical. It is for this reason that the university staff
perform risky operations and do not make sure appropriate preventive mechanisms are
in place. They simply use private life habits in the organization, i.e. they are not
sensitive enough to threats.
References
1. Andersen, R.: Inzynieria zabezpieczen. WNT, Warszawa (2005)
2. Lukatsky, A.: Wykrywanie wlaman i aktywna ochrona danych. Helion, Warszawa (2004)
3. Pakula, J.: Bezpieczny komputer. Czyli jak chronic sie przed hakerami, wirusami, spywarem,
spamami itd., Help (2005)
4. Pieprzyk, J., Hardjono, T., Seberry, J.: Teoria bezpieczenstwa systemow komputerowych.
Helion (2005)
5. Raport Ernst & Young: “Swiatowe badanie bezpieczenstwa informacji 2015”
6. Wisniewska, M.: Influence of organizational culture of a company on information security, In:
Lewandowski, J., Kopera, S., Krolikowski, J. (eds.) Innovation and knowledge in Innovative
Enterprise, Wydawnictwo Politechniki Lodzkiej. Series: Monographs, Lodz (2007)
7. Zalewski, M.: Cisza w sieci Praktyczny przewodnik po pasywnym rozpoznawaniu i atakach
posrednich. Helion, Warszawa (2005)
8. Wisniewska, M., Wisniewski, Z.: The relationship between knowledge security and the
propagation of innovation. Adv. Intell. Syst. Comput. 783, 176–184 (2019)
Beyond Passwords: Enforcing Username
Security as the First Line of Defense
Fort Hays State University, 600 Park Street, Hays 67601, USA
[email protected], [email protected]
1 Introduction
Access control and digital authentication algorithms typically use a combination of two
types of information for authorizing a user: account identifier and password. In general,
the former serves the purpose of finding a specific resource (or user) in the system,
whereas the latter requires solving the challenge of knowing the secret word that grants
access to the specific resource. To this end, the password is compared against the string
associated with the identifier stored in the system. This method has been successfully
utilized for decades for individual user accounts (e.g., e-mail mailbox, social media,
and on-line banking accounts) as well as for accessing resources that can be shared
among multiple users (e.g., public wireless networks).
Unfortunately, in the recent years, cyberattacks have become increasingly fierce
thanks to the accessibility of information on the Internet, the availability of more
powerful computational and network resources, and the development of more invasive
techniques for threatening victims [1]. As an example, clusters of distributed devices,
2 Related Work
help protect credentials. Indeed, this might disrupt the dynamics of websites, such as,
social media, that use the account name as a nickname. Conversely, this opportunity
could be utilized in other types of resources: to avoid impacting convenience, it could
be given as an option, or requested ad hoc, whenever a major security concern arises.
4 Study
In our analysis, we evaluated the aspects involved in the security and convenience
trade-offs that are related to both system design and user behavior (i.e., Sects. 3.1, 3.2,
3.3, 3.4, 3.5, and 3.6). Results are reported in Table 1. Specifically, the answers of G1
were used as a gold standard to evaluate the difference in terms of perceived security
between the two groups.
Conversely, in regard to convenience, we aggregated the data of the two groups:
this skewed the results, though it had a negligible impact on the outcome of our
analysis. This was for a two-fold reason: (1) although there was statistical significance
(p = 0.05) between the groups for two of the considered dimensions, we found that
perception of convenience was in general independent from cybersecurity training;
also, (2) password generation and verification systems serve all users in the same
fashion without considering their individual background or awareness.
Table 1. Results are shown in terms of perceived security (Sec), perceived convenience (Conv),
and occurrence (Occ). As for the former two dimensions, a Likert scale (1 being the least
secure/convenient) was utilized to collect responses; values regarding occurrence were directly
collected in percentages. Data are reported separately for the control group (G1) and respondents
with no cybersecurity training (G2).
Security factor Sec. G1 Sec. G2 Conv. Occ. Occ. G2
G1
Show the account name in clear 2.12 3.61 4.23 98.75% 100.00%
(Sect. 3.1)
Use login information as account 3.34 3.98 4.15 87.33% 78.22%
name (Sect. 3.2)
Give real-time feedback on 3.53 4.62 3.11 8.44% 11.33%
duplicate username (Sect. 3.3)
Use the e-mail address as account 2.44 4.32 4.54 68.80% 72.10%
name (Sect. 3.4)
Use personal information as 1.98 3.64 4.12 74.12% 79.78%
account name (Sect. 3.5)
Reuse the same name for multiple 1.52 3.37 4.27 82.22% 86.93%
accounts (Sect. 3.6)
Nevertheless, having the option of changing account name was accepted by the
majority of respondents, and it received approximately 89% of preferences. Further-
more, users perceived that they would not be affected by removing real-time feedback
about existing user names (i.e., Sect. 3.3), as this option was the most favored (a total
of 90% on average among the two groups). However, this might be caused by lack of
awareness of the consequences.
In general, users without cybersecurity training perceived current username prac-
tices as secure, and their responses show that they are more worried about their
Fig. 1. Perceived security (line) and willingness to use (bars). Data about willingness to adopt
(shown in Table 1) were converted from percentage to a Likert-equivalent to improve
visualization. Labels on the X axis correspond to the security items outlined in Sect. 3 and
described in Table 1.
56 T. Fandakly and N. Caporusso
passwords. This is consistent with the literature and with current practices. However,
account leak due to real-time feedback was perceived as resulting in minimal risk by
both G1 and G2. This practice was reported in 9.89% of cases, only.
Indeed, users know that reusing accounts that include personal information
diminishes security, though both G1 and G2 reported doing it on average in 78.17%
and 83.36% of the cases, respectively. This is consistent with findings in the literature
[18], which report that users are aware of malpractices, but they prefer to more con-
venient options because they think they will not be affected by an attack. We did not
detail account type because we found no statistically significant difference in perceived
convenience and security.
6 Conclusion
References
1. Caporusso, N., Chea, S., Abukhaled, R.: A game-theoretical model of ransomware. In:
International Conference on Applied Human Factors and Ergonomics, pp. 69–78. Springer,
Cham, July 2018. https://doi.org/10.1007/978-3-319-94782-2_7
2. Dev, J.A.: Usage of botnets for high speed MD5 hash cracking. In: Third International
Conference on Innovative Computing Technology (INTECH 2013), pp. 314–320. IEEE,
August 2013
3. Brumen, B., Taneski, V.: Moore’s curse on textual passwords. In: 2015 28th International
Convention on Information and Communication Technology, Electronics and Micro-
electronics (MIPRO) (2015). https://doi.org/10.1109/MIPRO.2015.7160486
4. National Institute of Standards and Technology Special Publication 800-63B., p. 78, June
2017. https://doi.org/10.6028/NIST.SP.800-63b
5. Onaolapo, J., Mariconti, E., Stringhini, G.: What happens after you are pwnd: understanding
the use of leaked webmail credentials in the wild. In: Proceedings of the 2016 Internet
Measurement Conference, pp. 65–79. ACM, November 2016
6. Lenig, S., Caporusso, N.: Minecrafting virtual education. In: International Conference on
Applied Human Factors and Ergonomics, pp. 275–282. Springer, Cham (2018). https://doi.
org/10.1007/978-3-319-94619-1_27
7. Hunt, T.: Password reuse, credential stuffing and another billion records in have i been
pwned, May 2017. https://www.troyhunt.com/password-reuse-credential-stuffing-and-
another-1-billion-records-in-have-i-been-pwned/. Accessed 31 Jan 2018
8. Stainbrook, M., Caporusso, N.: Convenience or strength? Aiding optimal strategies in
password generation. In: International Conference on Applied Human Factors and
Ergonomics, pp. 23–32. Springer, Cham, July 2018. https://doi.org/10.1007/978-3-319-
94782-2_3
9. Basta, A.: Computer Security and Penetration Testing, 2nd edn. Cengage Learning.
VitalBook file (2015). Accessed 8 Aug 2013
10. Shi, Y.: A method of discriminating user’s identity similarity based on username feature
greedy matching. Paper Presented at the 2018 2nd International Conference on Cryptog-
raphy, Security, and Privacy, March 2018. https://doi.org/10.1145/3199478.3199512
11. Wang, L., Li, Y., Sun, K.: Amnesia: a bilateral generative password manager. In: 2016 IEEE
36th International Conference on Distributed Computing Systems (ICDCS), pp. 313–322
(2016)
12. Jenkins, J.L., Grimes, M., Proudfoot, J., Lowry, P.B.: Improving password cybersecurity
through inexpensive and minimally invasive means: detecting and deterring password reuse
through keystroke-dynamics monitoring and just-in-time warnings. Inf. Technol. Dev. 20(2),
196–213 (2013)
13. Perito, D., Castelluccia, C., Kaafar, M.A., Manils, P.: How unique and traceable are
usernames? In: Privacy Enhancing Technologies. Lecture Notes in Computer Science, pp. 1–
17 (2011). https://doi.org/10.1007/978-3-642-22263-4_1
14. Xiao, X., Caporusso, N.: Comparative evaluation of cyber migration factors in the current
social media landscape. In: 2018 6th International Conference on Future Internet of Things
and Cloud Workshops (FiCloudW), pp. 102–107. IEEE, August 2018. https://doi.org/10.
1109/W-FiCloud.2018.00022
15. Thomas, K., Li, F., Zand, A., Barrett, J., Ranieri, J., Invernizzi, L., Bursztein, E.: Data
breaches, phishing, or malware? Understanding the risks of stolen credentials. Paper
Presented at the 2017 ACM SIGSAC Conference on Computer and Communications
Security, October 2017. https://doi.org/10.1145/3133956.3134067
58 T. Fandakly and N. Caporusso
16. Caporusso, N., Stainbrook, M.: Comparative evaluation of security and convenience trade-
offs in password generation aiding systems. In: International Conference on Applied Human
Factors and Ergonomics. Springer, July 2019. (to be published)
17. Johansson, J.M., Brezinski, D.I., Hamer, K.L.: U.S. Patent No. US13277423, U.S. Patent
and Trademark Office, Washington, D.C. (2011)
18. Tam, L., Glassman, M., Vandenwauver, M.: The psychology of password management: a
tradeoff between security and convenience. Behav. Inf. Technol. 29(3), 233–244 (2010).
https://doi.org/10.1080/01449290903121386
19. Bošnjak, L., Brumen, B.: Improving the evaluation of shoulder surfing attacks. In:
Proceedings of the 8th International Conference on Web Intelligence, Mining and Semantics
(2018). https://doi.org/10.1145/3227609.3227687
20. Bevilacqua, V.: Retinal fundus biometric analysis for personal identifications. In:
International Conference on Intelligent Computing, pp. 1229–1237. Springer, Heidelberg,
September 2008
Social Engineering and the Value of Data:
The Need of Specific Awareness Programs
1 Introduction
Cybersecurity is a hotly debated topic all over the world, and the protection of infor-
mation is a priority for institutions, companies and individuals. A data breach can have
a high financial impact on a company, considering that in the range of 1 million to 50
million records lost, breaches can cost companies between $40 million and $350
million respectively [1]. In addition, companies have also to consider other significant
consequences for their business, such as the loss of intellectual property and reputa-
tional damage [2].
Cyber threats have been growing over the last few years and they are going to be
based on the exploitation of new opportunities.
On the one hand, security international reports stress the impact of the old cyber
threats, such as ransomware, phishing, spear phishing, data breaches (e.g. [3–5]).
Moreover, they highlight the importance of human factor, since mail and phishing
represent the primary malware infection vector [3] while social engineering is a critical
launchpad for email attacks [5].
On the other hand, new threats are made possible by the application of Internet of
Things and Artificial Intelligence [6]; furthermore, these technologies can strengthen
the existing threats, such as improving the frequency of phishing attacks.
Notwithstanding the fact that more and more innovative technical solutions are
available on the market to provide protection to companies and institutions, the
problem of cybersecurity is far from being solved.
The role of human factor in cybersecurity is a fundamental topic to gain a better
defense against cyber-attacks. Many authors indeed stress the importance of adopting a
holistic approach, given that cyber defense cannot be considered only from a technical
perspective but requires also a human-social viewpoint (e.g. [7–9]).
This paper is focused on workers’ perception of cyber-attacks based on social
engineering (SE), which is a method using psychological principles to manipulate
people’s perception to gain their confidence and lead them to disclose sensitive
information or to do something else (e.g. opening an e-mail attachment), for the
benefits of those who use these strategies (e.g. [8, 10]). SE is a successful technique
because it exploits human nature bypassing technological measures [11]. In fact, as
reported in [10], «We, as human beings, are all vulnerable to being deceived because
people can misplace their trust if manipulated in certain ways».
SE can be used for several purposes and by different actors, targeting people
through information directly posted by Internet users. SE can be executed in different
forms. Phishing, a massive distribution of emails to solicit personal information, and
spear phishing, targeting victims individually, are a form of SE. Moreover, SE can
exploit physical devices (baiting), for example an infected USB stick left unattended in
order to be found and used by people, with the consequence of installing malware onto
the computer. Finally, SE can be executed by phone (vishing) to trick people or by
exploiting information collected during a face to face conversation. Even though the
actual modalities of execution can cause different reactions in people [12], the focus of
SE is the social interaction.
2 Methodology
The study has involved 212 employees belonging to companies operating in the service
sector (94 in company X, and 118 in company Y). In each company, we have carried
out a cybersecurity awareness project aimed at the building of security culture. We used
an interactive approach to actively involve participants and discuss with them security
problems, and how to manage them.
More specifically, within each project we gathered 3–4 groups belonging to the
same company for a half-day workshop where we tackled some of the most common
security risk scenarios related to human behavior (e.g. choosing secure password, using
unsecure wi-fi services). We repeated this half-day workshop with different sets of
groups until all involved employees had attended. There were in total 13 groups for
company X and 16 for company Y, with an average of 7 per group.
Social Engineering and the Value of Data 61
In each workshop, group participants were presented with the list of considered
security risk scenarios and were asked to assign a mark to the credibility of each of
them (i.e., how plausible the scenario is) and to its success probability, using a scale
from 1 (low) to 5 (high).
At the beginning of each workshop we explained, to all groups present, each of
these security risk scenarios, by showing videos in the public domain or short excerpts
from well-known movies depicting the specific scenario and by illustrating real life
examples of them (e.g. actual phishing emails). Subsequently, groups split and each of
them separately discussed the presented scenarios, in order to estimate its credibility
and success probability in the light of their personal experience, both in business and in
private life.
After each group internally discussed and provided a consensus evaluation on both
the credibility and the success probability of the scenarios, we united all groups
together and a representative from each of them declared their conclusion. Next, we
conducted a discussion and a comparison among all participants in that workshop of
the various conclusions. Finally, we trained participants on the best behavioral prac-
tices to manage the presented security risk scenarios.
Some of these security risk scenarios were based on social attacks and engineering
techniques (e.g. phishing), still a relevant problem given that social attacks are very
frequent and can compromise data, secrets, and credentials [4]. The security risk
scenarios discussed in the paper are the following:
• Receiving emails asking for data or to perform some action (Phishing and spear
phishing)
• Receiving a phone call asking for information (Vishing)
• USB dropped in obvious places to employees (USB baiting)
• Face to face conversation
Note that the first three above listed scenarios refer to situations that intrinsically
constitute direct risk scenarios, in the sense that they directly lead to jeopardize
valuable assets. On the other side the last scenario describes a situation where there is
not an immediate danger but the consequences of careless behaviors may provide a
social engineering attacker with the information on which to successfully carry out the
above three scenarios.
We now report and discuss the main outcomes of our study related to the above listed
scenarios, also in the light of the different situations existing in companies X and Y.
In Figs. 1 and 2 we compare credibility and success probability results obtained in
each of the two companies. Reported numbers are, for each risk scenario, the average
across all groups involved of the consensus evaluation provided by each group.
As you can see, the success probability has an average mark slightly lower than
credibility in all scenarios apart from “Vishing” in both companies. This scenario refers
to a kind of interaction where people are naturally aware of the risk of being uncon-
sciously manipulated by astute people. Even without training or previous experience, it
62 I. Corradini and E. Nardelli
appears highly plausible to many that an able and empathic speaker can persuade others
during a phone conversation.
Also, scenario “USB Baiting” has in both companies the lower mark, most prob-
ably because the specific situation where a memory stick is dropped in obvious places
for employees is not a common happening. Moreover, it depends on the security policy
adopted by organizations, given that the use of a USB stick could be prohibited.
Finally, scenario “Face to face conversation” has received the highest mark in one
company and the second highest in the other one, which is reasonable given that face to
face interactions are common in any kind of job and people are aware that these
situations can be a very good opportunity to collect sensitive information.
In Figs. 3 and 4 we present the same data but arranged to compare the situation
between the two companies.
Figure 3 presents credibility marks. You can see that, in general, employees in
Company Y are less convinced by the plausibility of the presented risk scenario than in
Company Y. This may be explained by the fact that company X has been working on a
security culture project for a few years and their employees have been participating in
Social Engineering and the Value of Data 63
specific training sessions, while company Y is at its first experience. Moreover, most
participants of the company Y are not aware that an email can be sent to targeted
individuals such as spear phishing, as well they were not aware of the psychological
aspects of these security risks.
Moreover, as reported by periodical reports on the most common types of cyber-
security threats and cited above, phishing is actually the preferred vehicle for SE
attacks.
For what regards the success probability, whose comparison of the marks between
the two companies is shown in Fig. 4, there is no published data – to the best of
authors’ knowledge – about the actual success rate of the various threats. Annual
cybersecurity reports usually provide indications on the top threats and whether a threat
has become more or less common compared to the previous year [4, 5, 15]. This is
understandable since data about failed attacks are usually not disclosed, while suc-
cessful attacks cannot be usually hidden, for both their visible consequences and data
protection laws (e.g. GDPR, General Data Protection Regulation) that requires com-
panies to notify an authority data breaches when they occur.
64 I. Corradini and E. Nardelli
Another important aspect concerns data protection related to the use of social
media, which has been tackled during the discussion of the “face to face conversation”
scenario, by relating what happens on social media to what happens in a person
interaction. Discussing their behavior on social media, employees tend to minimize the
risks associated with certain behaviors. From their words it emerged that while com-
pany’s data protection is considered fundamental to business, their awareness of the
value of personal information is not so high: an often repeated comment was “I have
nothing to hide”, while in reality each person has some information to protect. This
dichotomy between attitude and behavior concerning privacy, which emerged with
higher frequency in Company Y, is well-known in literature as the privacy paradox
(e.g. [13, 14]).
4 Conclusions
It is clear that while digital technology is spreading everywhere security risks are
growing and have to be seriously tackled. Criminals tend to exploit every vulnerability
they can find; in addition, they will be able to exploit the advantages of Artificial
Intelligence and Internet of Things.
If technical solutions are adequate to solve technical problems, they are inappro-
priate to manage security cyber threats related to human nature based on social engi-
neering technique, e.g. phishing and spear phishing attacks. Hence, companies have to
adopt a holistic approach, able to include and balance “People, Process and Technol-
ogy” [15].
The lack of security awareness represents a vulnerability for every organization,
making SE attacks easier to carry out. Hence, people using digital technologies have to
be more and more aware of the risks involved with their use. In fact, even though
cybersecurity is considered by governments and institutions as a priority, the actual
behavior of people represents a challenge for any organization [16].
Therefore, building a cybersecurity culture in organizations [17, 18] is the best way
to develop and reinforce effective security practices [19].
In this paper we have described the outcome of a study involving 212 employees,
belonging to two companies in the service sector, who participated to a cybersecurity
awareness project aimed at the building of a security culture within the organization.
Employees had to evaluate the credibility and the success probability of each security
risk scenario presented.
In one company the project was carried out for the first time, while in the other
people had already participated in cybersecurity awareness training sessions. The
analysis shows that people in the latter company have a better comprehension of risks
related to the use of digital technologies.
Our study therefore provides support for the fact that without adding people to a
company defense arsenal, effectiveness of its cybersecurity is weakened. This is in line
with recommendations of recent cybersecurity reports [3, 5].
Social Engineering and the Value of Data 65
References
1. Ponemon Institute: Cost of a Data Breach Study: Global Overview (2018). https://
databreachcalculator.mybluemix.net/assets/2018_Global_Cost_of_a_Data_Breach_Report.
pdf
2. Allianz: Allianz Risk Barometer. Top Business Risks for (2018). https://www.agcs.allianz.
com/assets/PDFs/Reports/Allianz_Risk_Barometer_2018_EN.pdf
3. ENISA: Threat Landscape Report. 15 Top Cyberthreats and Trends (2018). https://www.
enisa.europa.eu/publications/enisa-threat-landscape-report-2018
4. Verizon: Data Breach Investigation Report (2018). https://enterprise.verizon.com/resources/
reports/DBIR_2018_Report_execsummary.pdf
5. CISCO, Cisco 2018 Annual Security Report (2018). https://www.cisco.com/c/dam/m/
digital/elq-cmcglobal/witb/acr2018/acr2018final.pdf
6. Brundage, M., Avin, S., Clark, J., et al.: The malicious use of artificial intelligence:
forecasting, prevention, and mitigation (2018). https://arxiv.org/abs/1802.07228
7. Schultz, E.: The human factor in security. Comput. Secur. 24(6), 425–426 (2005)
8. Corradini, I.: Human factors in hybrid threats: the need for an integrated view. In: Zorzino,
G., et al. (eds.) Hybrid Cyberwarfare and The Evolution of Aerospace Power: Risks and
Opportunities, pp. 85–96, CESMA (2017)
9. Ki-Aries, D., Faily, S.: Persona-centred information security awareness. Comput. Secur. 70,
663–674 (2017)
10. Mitnick, K.D., Simon, W.L.: The Art of Deception: Controlling the Human Element of
Security. Wiley, New York (2002)
11. Schneier, B.: Secrets and Lies. Wiley, New York (2000)
12. Bullée, J.W.H., Montoya, L., Pieters, W., Junger, M., Hartel, P.: On the anatomy of social
engineering attacks: a literature-based dissection of successful attacks. J. Invest. Psychol.
Offender Profiling 15(1), 20–45 (2018)
13. Barnes, S.: A privacy paradox: social networking in the United States. First Monday, 11(9)
(2006). https://firstmonday.org/article/view/1394/1312_2
14. Barth, S., de Jong, M.D.T.: The privacy paradox: investigating discrepancies between
expressed privacy concerns and actual online behavior – a systematic literature review.
Telematics Inform. 34(7), 1038–1058 (2017)
15. Schneier, B.: https://www.schneier.com/blog/archives/2013/01/people_process.html
16. De Bruijn, H., Janssen, M.: Building cybersecurity awareness: the need for evidence-based
framing strategies. Gov. Inf. Q. 34, 1–7 (2017)
17. Enisa: Cyber Security Culture in organizations (2018). https://www.enisa.europa.eu/
publications/cyber-security-culture-in-organisations
18. Corradini, I., Nardelli, E.: Building organizational risk culture in cyber security: the role of
human factors. In: AHFE 2018, pp. 193–202. Springer, Cham (2018)
19. Wilson, M., Hash, J.: Building an information technology security awareness and training
program. NIST Special Publication 800-50, USA (2003)
Awareness and Cyber-Physical Security
Human Centered Cyber Situation Awareness
1 Introduction
Distribution Statement A. Approved for public release. Distribution is unlimited. This material is
based upon work supported under Air Force Contract No. FA8702-15-D-0001. Any opinions,
findings, conclusions or recommendations expressed in this material are those of the author(s) and
do not necessarily reflect the views of the U.S. Air Force.
This is a U.S. government work and not under copyright protection in the U.S.;
foreign copyright protection may apply 2020
T. Ahram and W. Karwowski (Eds.): AHFE 2019, AISC 960, pp. 69–78, 2020.
https://doi.org/10.1007/978-3-030-20488-4_7
70 V. Mancuso et al.
including all factors affecting friendly and adversary forces”, and generalizes to three
aspects of awareness, Network Awareness, Mission Awareness, and Threat Aware-
ness. These areas of awareness can be further divided based on the activities of friendly
(blue), neutral (gray) and adversary (red) forces and their capabilities across the
spectrum of conflict. Cyber SA for cyber defense has been described as containing
seven critical aspects of awareness; the current situation, the impact of the attack, the
evolution of the situation, adversary behavior, the causation of the situation, the quality
of the information, and the plausible futures of the current situation [9].
Unlike many of the environments in which SA has traditionally been studied,
cyberspace has additional complexities due to its dynamic and intangible nature.
Because the environment itself does not have a physical form with meaning, cyber-
security analysts often rely on technology to form, maintain and utilize their Cyber SA,
often referred to as capabilities. Cyber SA capabilities produce a significant amount of
data about the environment and its current state. If not curated and designed correctly,
this leads to a gap between the available data and the data that is necessary in the
current task, also known as the decision-making gap [10]. Rather than focusing on
delivering a deluge of data, Cyber SA capabilities should focus on understanding the
situation, and providing the necessary information in a way that is cognitively usable
and facilitates the human decision-making process.
To best augment SA and thus human performance, success capabilities must take
into account the people, processes, and technology needed. Previous models of SA
suggest that it is a human cognitive state that can be assisted by technology, and
distributed across organizations and teams that have concrete tasks to complete.
Therefore, depending on their job role, organization and mission, the requisite SA
information and processes around them, will vary across the different user groups in
cyber operations. From a capability perspective, this implies that Cyber SA tools must
provide accurate, timely, comprehensive and actionable information for a wide spec-
trum of possibilities, user groups, and tasks. When developing capabilities, this creates
two necessary questions that must be answered for each user group, (1) How do you
characterize different user groups across cyber operations and (2) What are the
Cyber SA information requirements for each of those user groups.
1.1 Purpose
The purpose of this research was to develop a holistic view of Cyber SA of Senior
Leaders across different organizations in cybersecurity. We leveraged a card sort
methodology [11, 12] to extract the critical Cyber SA requirements. The card sort task
provided us with data on how Senior Leaders prioritize information allowing us to
build a hierarchy of high- and low- priority Cyber SA information, as well as uncover
several shared schemas that can be used to drive capability development and serve as
boundary objects for collaboration across mission partners. The goal of this paper is to
not only provide empirical direction towards the development of Cyber SA capabilities
focused on specific information requirements, but to demonstrate an alternate
methodology for user-centered design in developing Cyber SA technologies.
Human Centered Cyber Situation Awareness 71
2 Method
2.1 Materials
Using the [Network, Mission, Threat] [Blue, Grey, Red] paradigms, we developed a
set of 69 Cyber SA information categories that were refined amongst the research team
and other Subject Matter Experts. Table 1 shows the general categories of cards across
the 3 3 paradigm space. It is important to note that these Cyber SA information
categories, were designed to represent the information needs of Senior Leaders, and not
necessarily the entire problem space of Cyber SA, thus the strong skew towards Blue
and Mission related information.
Table 1. Categories of cards within [Network, Mission, Threat] [Blue, Grey, Red] paradigm
Red Grey Blue
Network Adversary Key Partner Key Terrain and Blue Key Terrain, Security
Cyber Terrain Security Compliance Compliance and Current
(1 Card) (2 Cards) Cyber/Network Activity
(17 Cards)
Mission Top Threat Partner Missions and Blue Mission, Staffing, and
Groups/Actors Associated Risk, Relevant Readiness
and Associated Geo-Political News (26 Cards)
TTPs (5 Cards)
(2 Cards)
Threat Known Threats Top Threat Activity, Top Threat Activity
and Vulnerabilities, and Incidents Affecting DoD & US, and
Vulnerabilities Affecting Partners, and this Organization and
of Adversaries Global Security and Attacks Indicators of Insider Threat
(2 Cards) (6 Cards) (8 Cards)
After crafting an initial set of Cyber SA categorizations, we piloted the study with
participants familiar with network security. During these pilots, we adjusted the lan-
guage of the cards, added new cards and refined the methodology to be used during
exercises with Senior Leaders. Once the final set of relevant Cyber SA information was
agreed upon, cards were printed on laminated 2.5 3.5 cards (the size of a playing
card) with a unique ID number in the upper left-hand corner for the purpose of logging
(Fig. 1).
72 V. Mancuso et al.
Fig. 1. Sample cards used for Cyber SA Elicitation Card Sort exercise
At the beginning of each activity Senior Leaders were briefed on the intended goal
of the study. They were then presented the set of cards and received instructions. Senior
Leaders were asked to sort the cards based on priority. They were asked to create three
piles consisting of high, medium, and low priority items. High priority items were
defined as information that is crucial to their job and that they want to see on a frequent
basis. Medium priority items were defined as information that they want to see but not
Human Centered Cyber Situation Awareness 73
on a frequent basis. Low priority items were defined as items they do not need to see
very often if at all. The size of each of the three groupings could be as large or as small
as they wanted. Blank cards were available during the exercise if they felt a type of
information was missing from the deck. Once all cards were sorted, the researcher
inquired into why the participant organized the cards the way they did. During this
phase, participants were given freedom to rearrange cards.
3 Results
Overall a high variance in the priority ratings across Senior Leaders was found. How-
ever, analyses including descriptive statistics, overall prioritization rankings, and inter-
and intra- organization disagreement in rankings showed some interesting patterns for
discussion.
Similarly, there were five items which were agreed as being of lowest priority, with
only one participant indicating that they were high, and 9 or more suggesting they were
low priority Table 4.
74 V. Mancuso et al.
4 Discussion
threat activity including Top Cyber Attacks and Indication of Insider Threat were
considered to be of highest priority. These items have direct impact to mission and
require immediate defensive steps to be taken to reduce vulnerabilities, block potential
threats, and take other remediation steps. All of the top priority items additionally
referred to the Senior Leader’s own organization and not partner organizations,
implying that Senior Leaders are inward thinking when it comes to their Cyber SA
information needs. The lowest priority items were system administrative items
including Security Compliance and Security Hygiene of partner organizations and
Status of Service Desk Tickets for their own organization. While important for securing
the network, this information is perhaps too granular for Senior Leaders as it has an
unclear direct association to ongoing missions.
While there was agreement on what constituted high and low priority, there was a
high level of ambiguity on what information was considered of medium priority. While
there are multiple potential reasons for this, one thing that was noticed while running
these card sort activities, was the variance in interpretation for many of these Cyber SA
concepts. During the sorting process, the researchers provided little explanation,
allowing the participants to interpret the items themselves. The highly variable inter-
pretations indicate that Senior Leaders have different expectations and understandings
of the meaning of many of these concepts.
Another interesting finding, was in the agreement on priorities across the organi-
zation. Based on their organizational missions, it was expected that there would be
some disagreement in how each organization prioritized information. Our findings
showed that Executive Leadership had a higher value on information about partner
organizations, while Intelligence and Operations had more interest in information on
the cyber posture and status. Executive leadership is responsible for maintaining
awareness of the enduring mission scope, in which the on-going missions are situated.
The interaction of enduring missions and the entire set of current ongoing missions
involve a broader scope, and have dependencies on external entities. This is evident in
them considering key terrain and associated risk to the partner organizations to have
higher priority. Missions often have external dependencies with partner organizations,
and the success or failure of those missions has dependencies with the partner’s cyber
status.
Unlike the Executive Leadership, Senior Leaders in Intelligence and Operations
often have a singular (or small number) mission focus. From this perspective, they have
a much more fine-grained viewpoint of individual mission dependencies within their
organization. Participants in Intelligence and Operations held higher regard to infor-
mation on the local organizations Security Compliance and Hygiene, and the Global
Security Alert Levels. This information is important as it provides Intelligence and
Operations with information on what the current readiness and posture of their current
organization is at any given point and time, and how that information may interact with
the security alert levels that are being reported across the broader security community.
This information is critical in helping them better focus their mission directives, and
achieve the necessary effectiveness.
The final discussion item was based on the disagreement within the organizations.
Within both Executive Leadership and Intelligence and Operations, there were a few
items that had a surprisingly high level of disagreement across participants. Within both
76 V. Mancuso et al.
organizations there was disagreement on the role of partners in their Cyber SA.
Executive Leadership, had high levels of disagreement on information on the Key Blue
Force Assets and Threats to Key Terrain of their partners, while Intelligence and
Operations had disagreement for Partners Top Cyber Incidents and the Risk to Key
Assets. These disagreements, resonate with the differing priorities discussed earlier. For
Executive Leadership, with a broader mission focus, information on Blue Force Assets
and Threats to Key Terrain is very important with participants agreeing that those items
were either high or medium priority for their own organization, however there was
disagreement in priority rankings for partners. Similarly, Intelligence and Operations
had disagreements on Cyber Incidents and Risk to Key Assets, however for those items
focused on their organization, there was much higher agreement and prioritization. We
feel that these findings indicate a continuing ambiguity on how partner infrastructure
and status impact the local mission.
5 Implications
The purpose of this work was the help better inform the development of capabilities for
Senior Leaders Cyber SA, in addition to practical implications, there are also numerous
research implications.
It was hypothesized before conducting the card sorts that a large number of
Cyber SA items would be rated of low priority and a small number would be rated of
high priority. However, on average Senior Leaders rated more than 30% of the
Cyber SA information items as High priority. Senior Leaders receive their awareness
through some sort of regular update brief (in military organizations these are often
referred to as Commanders Update Briefs, CUB). These update briefs are often a few
PowerPoint charts, and reflect what their staff think represents the Commanders Critical
Information Requirements (CCIRs). However, with the large amount of High priority
Cyber SA information it is unlikely they are receiving the awareness that they desire.
This implies that there is research needed to help develop capabilities to help Senior
Leaders better identify and communicate their CCIRs, and enable the communication
pipeline of those requirements from their staff.
One of the biggest surprises in the analysis was that there was very little overlap or
uniformity in the prioritizations, even within an organization. This finding indicates
that Cyber SA may not be generalizable rather, Cyber SA is driven by context,
specifically mission priorities, current needs and risk tolerance. From a research
standpoint, this implies that further work is needed in order to map how the context of
the mission and current status of an organization impacts Cyber SA needs. From a
design standpoint, Cyber SA capabilities must be designed to be composable, to allow
Senior Leaders and their staff to customize which information they see, and modify it
based on their current priorities, needs, and mission focus.
There was also high ambiguity in the rating of medium priority items, which maybe
due to different interpretations of Cyber SA concepts. This could result in diverging
mental models, and lack of common ground, which could become a barrier in col-
laboration across organizations. Further research on operationalizing and defining
fuzzy Cyber terms may help improve agreement, and thus collaboration across
Human Centered Cyber Situation Awareness 77
References
1. Endsley, M.R., Rodgers, M.D.: Situation awareness information requirements analysis for en
route air traffic control. In: Proceedings of the Human Factors and Ergonomics Society
Annual Meeting, pp. 71–75. SAGE Publications, Los Angeles (1994)
2. Kaber, D.B., Perry, C.M., Segall, N., McClernon, C.K., Prinzel III, L.J.: Situation awareness
implications of adaptive automation for information processing in an air traffic control-
related task. Int. J. Ind. Ergonom. 36(5), 447–462 (2006)
3. Rodgers, M.: Human Factors Impacts in Air Traffic Management. Routledge, London (2017)
4. Harrald, J., Jefferson, T.: Shared situational awareness in emergency management mitigation
and response. In: HICSS 2007 40th Annual Hawaii International Conference on System
Sciences, pp. 23–23. IEEE (2007)
5. McNeese, M.D., Connors, E.S., Jones, R.E., Terrell, I.S., Jefferson Jr, T., Brewer, I., Bains,
P.: Encountering computer-supported cooperative work via the living lab: application to
emergency crisis management. In: Proceedings of the 11th International Conference of
Human-Computer Interaction (2005)
6. McNeese, M.D., Mancuso, V.F., McNeese, N.J., Endsley, T., Forster, P.: Using the living
laboratory framework as a basis for understanding next-generation analyst work. Security,
and Sensing, Baltimore, Maryland, Paper presented at the SPIE Defense (2013)
7. Endsley, M.R.: Toward a theory of situation awareness in dynamic systems. Hum. Factors
37(1), 32–64 (1995)
8. Joint Publications, Joint Publications 3–12 Cyberspace Operations. http://www.jcs.mil/
Doctrine/Joint-Doctrine-Pubs/3-0-Operations-Series/ (2018)
9. Barford, P., et al.: Cyber SA: Situational awareness for cyber defense. In: Cyber Situational
Awareness, pp. 3–13. Springer, Boston (2010)
10. Endsley, M.R., Garland, D.: Theoretical underpinnings of situation awareness: a critical
review. In: Endsley, M.R., Garland, D. (eds.) Situation Awareness Analysis and Measure-
ment, pp. 3–32. Taylor & Francis, Mahwah (2000)
11. Asgharpour, F., Liu, D., Camp, L.J.: Mental models of security risks. In: International
Conference on Financial Cryptography and Data Security, pp. 367–377. Springer,
Heidelberg (2007)
12. Falks, A., Hyland, N.: Gaining user insight: a case study illustrating the card sort technique.
Coll. Res. Libr. 61(4), 349–357 (2000)
13. Fleiss, J.L.: Measuring nominal scale agreement among many raters. Psychol. Bull. 76(5),
378–382 (1971)
14. Star, S.L., Griesemer, J.R.: Institutional ecology, translations’ and boundary objects:
amateurs and professionals in Berkeley’s museum of vertebrate zoology, 1907-39. Soc.
Stud. Sci. 19(3), 387–420 (1989)
Over-the-Shoulder Attack Resistant Graphical
Authentication Schemes Impact
on Working Memory
1 Introduction
brute force attacks, but this also can make the password more difficult to remember. Any
next-generation authentication scheme needs to strike a balance between security
requirements and usability [4]. Popular graphical next-generation schemes involve the
use of pictures instead of alphanumeric characters. People can quickly and easily
remember pictures versus strings of letters, symbols, and numbers. However, this very
strength of graphical authentication presents a weakness for security. A casual attacker
can glance over-the-shoulder of a user and steal their credentials. A recent report sug-
gests that 81% of users are mindful of over-the-shoulder attacks in public places [3]. The
recognition of this security vulnerability has produced over a decade of possible
deterrents against casual bystanders. They can be classified as schemes that group
targets among distractors, translate them to another location, and disguise the appear-
ance of targets. These schemes can help prevent Over-the-Shoulder Attacks (OSA) [6].
These OSA resistant graphical authentication schemes show impressive passcode
retention. The previous literature over the last decade strongly suggests that visually
rich stimuli such as pictures facilitate memorability [7]. Further, images offer extra
sensory data (c.f., letter, numbers, and special characters), which helps encoding and
later retrieval (i.e., picture superiority effect; [8]). According to Still, Cain & Schuster
(2017), recognizing and selecting graphics is often easier than recalling and generating
a password [9]. This is similar to the effort required to recognize a correct answer
within a multiple-choice test compared with writing an essay. A multiple-choice
question provides the tester with retrieval cues that facilitate long-term memory access
[10]. According to Cain and Still (2018), participants were able to remember graphical
passcodes following a three-week delay; however, almost no one remembered their
alphanumeric passcode [5]. Even when users are provided only a few practice trials,
they do not forget system assigned graphical passcodes weeks later.
Even with a significant memory advantage, there are other factors to consider that
might harm ease of use. For instance, a complex login procedure could make
authentication difficult. More specifically the amount/type of working memory an
authentication system drains might also predict its perceived “ease of use.” Currently,
no one has examined the impact graphical authentication has on working memory. We
find this surprising given authentication is a secondary task (i.e., users simply want past
security to work towards a goal). Further, authentication is often required numerous
times throughout the day. The previous literature has focused on making authentication
schemes efficient [11] and recognizing standards [12]. However, the cost of authenti-
cation can go beyond simply inconveniencing users. It is likely that schemes that
consume our limited working memory resources could harm our ability to remember
primary task information (e.g., the reason you needed to authenticate in the first place).
Baddeley described working memory as a conscious workbench with a limited
capacity. Using the workbench is effortful and directed, but flexible as we assemble, or
encode, information into long-term memory. Working memory plays an important role
in interface interactions. Baddeley’s model of working memory is composed of three
main components: central executive, verbal (visuo-spatial sketchpad), and visual (ar-
ticulatory loop) [13]. The verbal storage is fleeting with a duration of approximately
2000 ms. However, rehearsal, like subvocalization, can maintain the memory from
being lost. Classic examples include remembering items in a list such as drink pref-
erences for an upcoming party. The spatial storage helps users resolve every
Over-the-Shoulder Attack 81
2 Method
2.1 Participants
Eighteen undergraduate volunteers participated for course credit. Our sample was
composed of 14 females, 15 native English speakers, and they reported having an
average of 7 h of daily computer use (SD = 3.05). Two participants had to be excluded
due to data collection errors, which resulted in missing data or placement in the wrong
condition. We needed participants who were able to demonstrate an ability to
authenticate. Following removal of participants with a login proficiency below 80%,
only nine participants remained; three per authentication type. The remaining 9 par-
ticipants were able to successfully authenticate; (UYI: M = 99; SD = 1.73), (WYS-
WYE: M = 92; SD = 8.54), (CHC: M = 89; SD = 3.46). Clearly, our brief graphical
authentication training procedure was probably not sufficient; (CHC: M = 73; SD =
6.81), (WYSWYE: M = 69; SD = 10.15).
2.3 Procedure
Participants were run individually. They were seated in front of two desktop computers.
On one computer Paradigm© presented the graphical schemes, and on the other E-
Prime© presented the working memory task. Each participant interacted with a single
graphical scheme. However, every participant received all the different types of working
Over-the-Shoulder Attack 83
memory load. The working memory type presentation order was not completely
counterbalanced. Before each authentication trial, the working memory task presented
items to be remembered, and after each trial the participants typed the items to the best of
their ability. Participants were provided with instructions, 20 authentication scheme
practice trials, and three working memory task practice trials. The experimenter provide
feedback on whether participants had correctly authenticated or not on every trial.
3 Results
All the statistical tests used an alpha level of .05. Clearly, the overall numerical trends
show the negative impact authentication has on information held in working memory.
To explore our manipulations, a repeated measures ANOVA examining working
memory type (verbal, spatial, central executive) as a function of authentication scheme
(between-subject factor: UYI, CHC, WYSWYE) revealed significant main effects on
working memory accuracy. The analysis revealed a within-subjects main effect of
Working Memory Type, F(2,12) = 16.84, p < .001, n2p = .70. And, the between-
subjects main effect of Authentication Scheme, F(2,6) = 9.71, p = .013, n2p = .764.
However, the interaction between Working Memory Type X Authentication
Scheme was not found to be significant, F(4,12) = .601, p = .669, n2p = .05. Post-hoc
tests were conducted using Bonferroni with adjusted alpha levels to avoid Type I error
inflation across multiple comparisons.
Working Memory Type. Results indicated that Spatial working memory (M = .79,
SEM = .04) was best preserved compared to Verbal (M = .54, SEM = .06) and Central
Executive (M = .34, SEM = .07), p < .05. However, Verbal working memory and
Central Executive were not significant different from each other, p = .139.
Authentication Scheme. It was revealed that the CHC and WYSWYE authentication
schemes were the only schemes to be statistically different. Specifically, the CHC
(M = .67, SEM = .04) had less negative impact on working memory accuracy compared
with the WYSWYE scheme (M = .44, SEM = .08), p = .014, d = 1.466 (see Fig. 1).
WYSWYE
CHC
UYI
4 General Discussion
Authentication schemes must afford us with easy access to our data, while maintaining
their security. This is a difficult balancing act between user and technical requirements.
The conventional alphanumeric scheme has quickly evolved into a soft target.
Attackers know passwords are being reused and contain personal information. In an
attempt to strengthen passwords, the technical requirements for a strong password are
increased. This makes remembering a password difficult. Therefore, authentication
developers have been searching for a next-generation solution. One popular solution
proposed in the literature is graphical authentication.
These schemes offer users with superior memorability compared to password-based
schemes. However, a graphical authentication system usually requires more complex
interactions than simply typing in a password. For example, a user might be asked to
search for an icon among many decoys. It is recognized logging in with a graphical
authentication scheme is less efficient compared with conventional schemes (e.g.,
password or PIN). This could annoy users given they need to login numerous times
throughout the day. However, harming working memory can go beyond subjective
dissatisfaction. Surprisingly, no one has examined the impact of graphical authenti-
cation on the working memory system. Stealing limited working memory resources
from the primary task to authenticate might harm our primary task performance.
This pilot study found that working memory performance varies by graphical
authentication scheme. Clearly, authentication interactions influence the user’s ability
to remember information being held in working memory for the primary task. We
believe that a larger sample size will reveal additional significant differences between
authentication schemes given the medium to large effect sizes (i.e., range = .66–.81).
Future work needs to explore the interaction between working memory types and
authentication schemes. We suggest a follow-up study, increases the number of par-
ticipants in the study and introduces a low-load and password conditions. This will
allow researchers to explore the actual impact of working memory type independent of
task difficulty.
Acknowledgments. We thank Paige Duplantis, Lauren Tiller, and Ayobami Fakulujo for their
assistance collecting data.
References
1. Grawemeyer, B., Johnson, H.: Using and managing multiple passwords: a week to a view.
Interact. Comput. 23, 256–267 (2011)
2. Zviran, M., Haga, W.J.: Password security: an empirical study. J. Man. Info. Sys. 15, 161–
185 (1999)
3. Cain, A.A., Edwards, M.E., Still, J.D.: An exploratory study of cyber hygiene behaviors and
knowledge. J. Info. Secur. App. 42, 36–45 (2018)
4. Still, J.D.: Cybersecurity needs you! ACM Interact. (May + June: Feature). 23, 54–58 (2016)
5. Cain, A.A., Still, J.D.: Usability comparison of over-the-shoulder attack resistant authen-
tication schemes. J. Usab. Stud. 13, 196–219 (2018)
Over-the-Shoulder Attack 85
6. Cain, A.A., Werner, S., Still, J.D.: Graphical authentication resistance to over-the-shoulder-
attacks. In: Proceedings CHI Conference Extended Abstracts, pp. 2416–2422 (2017)
7. Biddle, R., Chiasson, S., Van Oorschot, P.C.: Graphical passwords: learning from the first
twelve years. ACM Comp. Sur. (CSUR) 44, 1–25 (2012)
8. Mintzer, M.Z., Snodgrass, J.G.: The picture superiority effect: support for the distinctiveness
model. Amer. J. Psyc. 112, 113–146 (1999)
9. Still, J.D., Cain, A., Schuster, D.: Human-centered authentication guidelines. Info.
Comp. Sec. 25, 437–453 (2017)
10. Tulving, E., Thomson, D.M.: Encoding specificity and retrieval processes in episodic
memory. Psyc. Rev. 80, 352–373 (1973)
11. Werner, S., Hauck, C., Masingale, M.: Password entry times for recognition-based graphical
passwords. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 60, 755–759 (2016)
12. Braz, C., Robert, J.: Security and usability: the case of the user authentication methods. In:
Proceedings of the 18th International Conference on Association Francophone d’Interaction
Homme-Machine, 199–203 (2006)
13. Baddeley, A.: Working memory. Science 255, 556–559 (1992)
14. Logie, R.H.: Retiring the central executive. Q. J. Exp. Psychol. (2016). advance online
publication
15. Wiedenbeck, S., Waters, J., Sobrado, L., Birget, J. C.: Design and evaluation of a shoulder-
surfing resistant graphical password scheme. In: Proceedings of the Working Conference on
Advanced Visual Interfaces, pp. 177–184 (2006)
16. Hayashi, E., Dhamija, R., Christin, N., Perrig, A.: Use your illusion: secure authentication
usable anywhere. In: Proceedings of the 4th Symposium on Usable Privacy and Security,
pp. 35–45 (2008)
17. Khot, R.A., Kumaraguru, P., Srinathan, K.: WYSWYE: shoulder surfing defense for
recognition based graphical passwords. In: Proceedings of the 24th Australian CHI
Conference, pp. 285–294 (2012)
18. Ankush, D.A., Husain, S.S.: Authentication scheme for shoulder surfing using graphical and
pair based scheme. Intern. J. Adv. Res. Comp. Sci. Mang. Stud. 2, 161–166 (2014)
19. Behl, U., Bhat, D., Ubhaykar, N., Godbole, V., Kulkarni, S.: Multi-level scalable textual-
graphical password authentication scheme for web based applications. J. Electron. Commun.
3, 166–124 (2014)
20. Chen, Y.L., Ku, W.C., Yeh, Y.C., Liao, D.M.: A simple text-based shoulder surfing resistant
graphical password scheme. In: IEEE ISNE, pp. 161–164 (2013)
21. Joshuva, M., Rani, T.S., John, M.S.: Implementing CHC to counter shoulder surfing attack
in PassPoint–style graphical passwords. Intern. J. Adv. Net. App. 2, 906–910 (2011)
22. Kiran, T.S.R., Rao, K.S., Rao, M.K.: A novel graphical password scheme resistant to
peeping attack. Int. J. Comput. Sci. Inf. Technol. 3, 5051–5054 (2012)
23. Manjunath, G., Satheesh, K., Saranyadevi, C., Nithya, M.: Text-based shoulder surfing
resistant graphical password scheme. Intern. J. Comp. Sci. Info. Tech. 5, 2277–2280 (2014)
24. Rao, K., Yalamanchili, S.: Novel shoulder-surfing resistant authentication schemes using
text-graphical passwords. Int. J. Inf. Secur. 1, 163–170 (2012)
25. Vachaspati, P.S.V., Chakravarthy, A.S.N., Avadhani, P.S.: A novel soft computing
authentication scheme for textual and graphical passwords. Intern. J. Comp. App. 71, 42–
54 (2013)
26. Zhao, H., Li, X.: S3PAS: a scalable shoulder-surfing resistant textual-graphical password
authentication scheme. In: 21st AINAW, vol. 2, pp. 467–472 (2007)
27. Tiller, L., Cain, A., Potter, L., Still, J.D.: Graphical authentication schemes: balancing
amount of image distortion. In: Ahram, T., Nicholson, D. (eds.) Advances in Human Factors
in Cybersecurity, pp. 88–98 (2019)
86 J. D. Still and A. A. Cain
28. Cain, A.A., Still, J.D.: A rapid serial visual presentation method for graphical authentication.
In: Nicholson, D. (ed.) Advances in Human Factors Cybersecurity, pp. 3–11. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-41932-9_1
29. Gao, H., Guo, X., Chen, X., Wang, L., Liu, X.: Yagp: yet another graphical password
strategy. In: Computer Security Applications Conference, pp. 121–129 (2008)
30. Ghori, F., Abbasi, K.: Secure user authentication using graphical passwords. J. Ind. Stud.
Res. 11, 34–40 (2013)
31. Hui, L.T., Bashier, H.K., Hoe, L.S., Kwee, W.K., Sayeed, M.S.: A hybrid graphical
password scheme for high-end system. Aust. J. Bas. App. Sci. 8, 23–29 (2014)
32. Jenkins, R., McLachlan, J.L., Renaud, K.: Facelock: familiarity-based graphical authenti-
cation. Peer J. 2, 1–24 (2014)
33. Lin, D., Dunphy, P., Olivier, P., Yan, J.: Graphical passwords & qualitative spatial relations.
In: Proceedings of Symposium on Usable Privacy and Security, pp. 161–162 (2007)
34. Meng, Y., Li, W.: Enhancing click-draw based graphical passwords using multi-touch on
mobile phones. In: IFIP Conference, pp. 55–68 (2013)
35. Nicholson, J.: Design of a Multi-touch shoulder surfing resilient graphical password. B. Sci.
Info. Sys. (2009)
36. Sasamoto, H., Christin, N., Hayashi, E.: Undercover: authentication usable in front of prying
eyes. In: Proceedings of the SIGCHI Conference, pp. 183–192 (2008)
37. Yakovlev, V.A., Arkhipov, V.V.: User authentication based on the chess graphical password
scheme resistant to shoulder surfing. Auto. Con. Comp. Sci. 49, 803–812 (2015)
38. Zakaria, N.H., Griffiths, D., Brostoff, S., Yan, J.: Shoulder surfing defense for recall-based
graphical passwords. In: Proceedings of Seventh Symposium on Usable Privacy and
Security, pp. 6–18 (2011)
39. Bianchi, A., Oakley, I., Kim, H.: PassBYOP: bring your own picture for securing graphical
passwords. IEEE Trans. Hum. Mach. Syst. 46, 380–389 (2016)
40. Brostoff, S., Inglesant, P., Sasse, M.A.: Evaluating the usability and security of a graphical
one-time PIN system. In: Proceedings of the 24th BCS Interaction Specialist Conference,
pp. 88–97 (2010)
41. De Luca, A., Hertzschuch, K., Hussmann, H.: ColorPIN: securing PIN entry through indirect
input. In: Proceedings of the SIGCHI, pp. 1103–1106 (2010)
42. Gao, H., Liu, X., Dai, R., Wang, S., Chang, X.: Analysis and evaluation of the colorlogin
graphical password scheme. In: Fifth International Conference on Image and Graphics,
pp. 722–727 (2009)
43. Gupta, S., Sahni, S., Sabbu, P., Varma, S., Gangashetty, S.V.: Passblot: a highly scalable
graphical one time password system. Intern. J. Net. Sec. App. 4, 201–216 (2012)
44. Kawagoe, K., Sakaguchi, S., Sakon, Y., Huang, H.H.: Tag association based graphical
password using image feature matching. In: International Conference on Database Systems
for Advanced Applications, pp. 282–286 (2012)
45. Lashkari, A.H., Manaf, A.A., Masrom, M.: A secure recognition based graphical password
by watermarking. In: 11th International Conference on Computer and Information
Technology, pp. 164–170 (2011)
46. Perkovic, T., Cagalj, M., Rakic, N.: SSSL: shoulder surfing safe login. In: 17th International
Conference Software, Telecommunications & Computer Network, pp. 270–275 (2009)
47. Zangooei, T., Mansoori, M., Welch, I.: A hybrid recognition and recall based approach in
graphical passwords. In: Proceedings of the 24th Australian CHI Conference, pp. 665–673
(2012)
48. Still, J.D., Dark, V.J.: Examining working memory load and congruency effects on
affordances and conventions. Int. J. Hum Comput Stud. 68, 561–571 (2010)
Comparative Evaluation of Security
and Convenience Trade-Offs in Password
Generation Aiding Systems
Fort Hays State University, 600 Park Street, Hays 67601, USA
[email protected], [email protected]
Abstract. A strong password is considered the most important feature for the
security of any account credentials. In the last decades, several organizations
focused on improving its strength and produced awareness initiatives and
security guidelines on how to create and maintain secure passwords. However,
studies found that users perceive security and convenience as a trade-off, and they
often compromise password strength in favor of a key phrase that is easier to
remember and type. Therefore, nowadays websites and applications implement
password generation aiding systems (PGAS) that help, and even force, users to
create more secure passwords. Nowadays, several types of PGAS are available,
each implementing a different strategy for stimulating users in crating stronger
and more secure passwords. In this paper, we present the results of a study in
which we compared six different PGAS and evaluated their performance in terms
of security and convenience, with the aim of suggesting the system that has the
most beneficial trade-off depending on the type of application.
1 Introduction
In the last decade, novel access systems implementing sophisticated credentials, such
as, two-factor authentication and biometric identification [1], have increasingly been
utilized for protecting the security and privacy of information in devices and accounts.
Nevertheless, text-based passwords are still the most common authentication method
for accessing websites, electronic mailboxes, and other types of accounts (e.g., wireless
networks). However, the main limitation of passwords lies in the paradox of the trade-
off between security and convenience: strong and secure passwords typically are
inconvenient and difficult to remember [2]. As a result, research showed that users tend
to utilize a total of 3–6 passwords shared between multiple accounts, even if they adopt
a password manager; moreover, they prefer to create key phrases that are easy to
remember by including information that is meaningful to them (such as, important
dates and names) which, in turn, affects their security [3].
Password generation aiding systems (PGAS) have been developed to address the
issue and help users increase the strength of their credentials by enforcing specific
requirements (e.g., minimum length, presence of special symbols, or entropy) when
they create a new password. Although they provide users with feedback about security
score, research showed that users typically approach most PGAS as a checklist, and
they do not result in any security improvement beyond the lowest mandatory level [3].
Indeed, there are several types of PGAS and they have very different strategy, design,
and interface characteristics. However, only a few studies took into consideration the
performance implications of the different components of user experience in password
generation aiding systems, such as, interface, usability, and type and timeliness of
feedback.
2 Related Work
password feedback into account when changing their passwords on these sites.
Additionally, researchers found that providing a password meter for users when
changing passwords helps to influence users to create stronger passwords, opposed to
showing a password meter when first creating the account [7].
Password meters are currently active in many of Alexa’s 100 most visited global
sites. Out of the top 100 sites, 96 allow the ability for users to create a password and
account, out of the 96, 73% gave some sort of feedback to users on their passwords,
many of the sites used similar or the same password meter ranging from a bar meter to
checkmarks systems after meeting requirements [8]. These systems were dynamically
updated as the user types and considered the length, characters used, and occasionally
blacklisted words [10]. A large-scale study on the effectiveness of password meters
found that overall, password meters do change user behavior when interacting with
them. Researchers found that users seeing a password meter nudged users to create
longer passwords. Furthermore, findings suggested that the visual component of a
password meter did not lead to significant differences. Users presented with a dancing
bunny password meter reacted the same as those presented with traditional password
feedback; however, the combination of text and a visual component was an important
factor in the effectiveness. Additionally, the researchers found that users substantially
changed their behavior when they were presented with stringent meters that would
purposefully lower the security score of the password to motivate users to create a more
secure password. Moreover, the stringent meters did motivate users to change their
passwords; however, users also reported these meters to be annoying and frustrating to
use [9].
Most research agrees that password feedback can help improve user’s passwords
[6, 7, 9]; however, password meters with requirements may be viewed by some users as
a checklist. A study on the impact of feedback on password-creation behavior found
that some users would quit adding characters and length to their passwords after
fulfilling the password feedback requirements, such as: minimum of eight characters, at
least one upper and lower-case letter, etc. Researchers theorized that some users may
view the feedback as requirements and quit improving their password security after
fulfilling them as they give the user a feeling of finality from reaching the requirements.
Whereas, in situations where the password feedback did not have requirements, users
may not be aware that the requirements were met and add additional security such as
length and special characters. Another, theory the researchers had about why users may
stop adding security after meeting requirements is that they rely on the feedback system
for security. Therefore, users trust the feedback system, relying on it to help them create
a secure password. Ultimately, the researchers recommend password feedback systems
prompt users to continue adding security mechanisms to their passwords after fulfilling
password requirements [10].
Implementing password feedback systems and meters may help to improve user
passwords by giving explicit instructions and providing a visual representation, usually
in a horizontal bar that increases and fills as the security increases. Moreover, this does
not help improve the problem of users and poor password management. A large-scale
study of user-password interaction found that the average user had 6.5 passwords
shared between 3.9 sites [8]; another study found that 63.8% users reported using their
password elsewhere, despite being aware of security practices [7]. This illustrates the
90 M. Stainbrook and N. Caporusso
In this Section, we review the most common strategies utilized in PGAS for helping
users increase the security of their passwords: reactive techniques include suggesting
guidelines, enforcing requirements, and giving feedback, whereas proactive methods
automatically generate a secure key phrase for the user. They are detailed in Fig. 1.
Although their functionality might vary depending on the type of information in an
account, their general purpose is to improve the trade-off between security, that is,
effort to crack, and convenience (i.e., effort to generate and use).
prevents creating passwords that are below a specific strength and (2) it educates users
about security by having them practice.
4 Experimental Study
In this Section, we detail a study in which we compared the different strategies and
types of PGAS. Specifically, we evaluated the impact of human factors on the rela-
tionship between usability and performance in terms of security and convenience trade-
off. To this end, we designed a web-based account creation system that implements
the six different PGAS discussed in Sect. 3. We recruited a total of 115 subjects
(33 females and 82 males aged 28 ± 9). Participants were sampled from a population
having low to medium familiarity with IT (based on degree, background, time spent
working with a desktop computer, and number of accounts) to prevent results from
being biased by computer literacy.
In the experimental task, participants were directed to an Internet URL in which:
(1) they were presented with a sign-up form incorporating a password generation aiding
system selected at random and they were asked to create an account using a key phrase
they never utilized in the past and save the information using the method of their
choice; (2) they were redirected to an account activation page that had the purpose of
interrupting their task; (3) they were presented with a sign-in form and they were asked
to use their credentials. Finally, they were asked to fill a questionnaire in which they
evaluated the convenience and usability of PGAS in the sign-up phase and in signing
in. Participants were asked to realize the task six times (one for each type of system).
The order of PGAS was selected at random to prevent training effect. The experimental
software calculated the entropy of the password generated using each PGAS. More-
over, it acquired the time required for typing the key phrase at sign up (calculated from
the first character entered until the user left the password field) and for signing in
(calculated from page load until the user clicked the sign in button). The security
criteria were the same for all systems: minimum 8 characters long, at least a number
and an uppercase and a lowercase letter.
Comparative Evaluation of Security and Convenience Trade-Offs 93
Fig. 1. Different types of Password Generation Aiding Systems and strategies as described in
Sect. 3 and utilized in the study: showing password requirements (3.1) is the most basic approach
to security, which can be enforced using systems that require the specifications to be met (3.2).
Strength meters (3.3) score password robustness whereas systems based on time to crack (3.4)
measure the security of a password in terms of time required to guess it, which can be estimated
using length and entropy measurement. Also, non-conventional meters display the strength of the
chosen password as compared to other users (3.5). Finally, a few software and websites are
automatically generating secure passwords for the user when they register (3.6).
PGAS that suggest or require specific symbols resulted in the lowest overall score, as
they required 4.20 s on average to create and use, had the least entropy and length, and
were preferred last in terms of convenience. In line with findings in the literature,
subjects utilized them as a checklist. Conversely, password meters had better results,
though label-based systems had different outcome than meters based on time-to crack
and peer comparison. Specifically, the former resulted lasts in terms of users’ per-
ception even if they had better performance, both as calculated effort to create and use,
and as resulting entropy and length. Finally, PGAS that automatically generate a
password resulted in the lowest effort to create and use and in the best preference.
Although they require users to store the key phrase in password manager, they also
prevent reusing the same word across multiple accounts and, thus, might have an
94 M. Stainbrook and N. Caporusso
impact on the overall security of the user. Table 1 shows a summary of all the
experiment data. Figure 2 reports the effort to create and use the password, calculated
as seconds spent in the sign up and sign in phases. Figure 3 indicates password
strength, measured in terms of key phrase entropy and length. Perceived convenience
in account creation and access was recorded using a Likert scale (see Fig. 4). Overall,
users spent between 3 and 5 s on average to sign up and sign in. The entropy of
password ranged from weak to strong depending on the PGAS, whereas the average
length was 10 characters. Finally, systems were perceived similarly, though there is a
statistical difference between them.
Table 1. Experiment data acquired from the different PAGS described in Sect. 3: password
guidelines (3.1), password requirements (3.2), strength meter (3.3), time to crack indicator (3.4),
peer strength meter (3.5), and password generator (3.6).
PGAS 3.1 3.2 3.3 3.4 3.5 3.6
Time to sign 4.15 ± 0.85 4.24 ± 0.88 4.11 ± 0.88 4.36 ± 1.44 4.12 ± 1.42 0.25 ± 0.83
up (seconds)
Time to sign 4.42 ± 1.44 4.26 ± 1.33 3.78 ± 1.17 5.08 ± 1.68 4.91 ± 1.74 5.75 ± 3.06
in (seconds)
Entropy (bits) 39.62 ± 7.30 41.22 ± 8.13 47.26 ± 8.80 43.28 ± 9.17 49.24 ± 8.30 47.5 ± 8.71
Length 9.45 ± 1.05 9.47 ± 1.15 11.25 ± 2.01 11.15 ± 1.92 11.12 ± 2.03 9.11 ± 0.57
(characters)
Convenience 3.38 3.47 3.41 4.36 4.29 4.85
in sign up
(Likert)
Convenience 3.12 3.21 3.77 2.77 3.04 2.74
in sign in
(Likert)
Fig. 2. Effort to create and use a password, measured in seconds required to generate it and to
use it. PGAS that automatically generate a password (3.6) ranked best, though there was is a 5 s
difference between sign up and sign in phases. Standard label-based password meters are the
most efficient alternative.
Comparative Evaluation of Security and Convenience Trade-Offs 95
Fig. 3. Password strength calculated as entropy and length. The latter, which was initially
measured in characters, was converted to bits (8 bits per character) for visualization purposes.
Password meters resulted in the highest score of entropy and length. However, this was because
only a few subjects changed the automatically-generated password in (3.6), which is statistically
comparable to 3.3 in terms of entropy.
Fig. 4. Perceived convenience in the sign up and sign in phases, scored on a Likert scale.
Systems that automatically generate a password for the user (3.6) ranked best, though individual
steps differ significantly. Also, respondents indicated a preference for password meters over
system that suggest or enforce password requirements.
6 Conclusion
In this paper, we presented an overview of the main types of currently available PGAS,
we highlighted their differences, and we discussed the results of a comparative eval-
uation of their security performance and perceived usability. From our findings, we can
conclude that interface design strategies lead to better trade-off between password
96 M. Stainbrook and N. Caporusso
References
1. Bevilacqua, V., Cariello, L., Columbo, D., Daleno, D., Fabiano, M.D., Giannini, M.,
Mas-tronardi, G., Castellano, M.: Retinal fundus biometric analysis for personal identifi-
cations. In: International Conference on Intelligent Computing, pp. 1229–1237, September
2008
2. Bonneau, J., Herley, C., Van Oorschoto, P.C., Stajano, F.: Passwords and the evolution of
imperfect authentication. Commun. ACM 58(7), 78–87 (2015). https://doi.org/10.1145/
2699390
3. Stainbrook, M., Caporusso, N.: Convenience or strength? Aiding optimal strategies in
password generation. In: Proceedings of Advances in Human Factors in Cybersecurity,
AHFE 2018. Advances in Intelligent Systems and Computing, vol 782 (2018). https://doi.
org/10.1007/978-3-319-94782-2_3
4. Tam, L., Glassman, M., Vandenwauver, M.: The psychology of password management: a
trade-off between security and convenience. Behav. Inf. Technol. 29(3), 233–244 (2010).
https://doi.org/10.1080/01449290903121386
5. Ciampa, M.: A comparison of password feedback mechanisms and their impact on password
entropy. Inf. Manag. Comput. Secur. 21 (2013)
6. Dupuis, M., Khan, F.: Effects of peer feedback on password strength. In: APWG Symposium
on Electronic Crime Research, San Diego, CA, pp. 1–9 (2018). https://doi.org/10.1109/
ECRIME.2018.8376210
7. Egelman, S., Sotirakopoulos, A., Muslukhov, I., Beznosov, K., Herley, C.: Does my
password go up to eleven? The Impact of password meter on password selection. In:
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems,
pp. 2379–2388. ACM, New York (2013). https://doi.org/10.1145/2470654.2481329
8. Florêncio, D., Herley, C.: A large-scale study of web password habits. In: Proceedings of the
16th International Conference on the World Wide Web, pp. 657–666 ACM Press, New York
(2007)
9. Ur, B., et al.: How does your password measure up? The effect of strength meters on
password creation. In: Proceedings Security 2012, USENIX Association (2012)
10. Shay, R., et al.: A spoonful of sugar? The impact of guidance and feedback on password-
creation behavior. In: Proceedings of the 33rd Annual ACM Conference on Human Factors
in Computing Systems, pp. 2903–2912, April 2015
11. Caporusso, N., Chea, S., Abukhaled, R.: A game-theoretical model of ransomware. In:
International Conference on Applied Human Factors and Ergonomics, pp. 69–78. Springer,
Cham, July 2018. https://doi.org/10.1007/978-3-319-94782-2_7
12. Xiao, X., Caporusso, N.: Comparative evaluation of cyber migration factors in the current
social media landscape. In: 2018 6th International Conference on Future Internet of Things
and Cloud Workshops (FiCloudW), pp. 102–107. IEEE, August 2018. https://doi.org/10.
1109/W-FiCloud.2018.00022
13. Fandakly, T., Caporusso, N.: Beyond passwords: enforcing username security as the first line
of defense. In: International Conference on Applied Human Factors and Ergonomics.
Springer, July 2019 (to be published)
Perceiving Behavior of Cyber Malware
with Human-Machine Teaming
Abstract. Cyber malware has evolved from simple hacking programs to highly
sophisticated software engineering products. Human experts are in high demand
but are busy, expensive, and have difficulty searching through massive amount
of data to detect malware. In this paper, we develop algorithms for machines to
learn visual pattern recognition processes from human experts and then to map,
measure, attribute, and disrupt malware distribution networks. Our approach is
to combine visualization and machine vision for an intuitive discovery system
that includes visual ontology of textures, topological structures, traces, and
dynamics. The machine vision and learning algorithms are designed to analyze
texture patterns and search for similar topological dynamics. Compared to recent
human-machine teaming systems that use input from human experts for
supervised machine-learning, our approach uses fewer samples, i.e. less training,
and aims for novel discoveries through human-machine teaming.
1 Introduction
Over the past two decades, malicious software have evolved from simple hacking
programs to highly sophisticated software engineering products. Prevailing data-driven
machine-learning methods, such as signature recognition and behavior profiling are not
sufficient to discover and attribute the malware with advanced stealthy functions such
as polymorphism and distributed hosting. In addition, rapidly growing malware con-
tains many unknown species that, given a lack of training samples, makes traditional
machine learning algorithms insufficient. Human experts are still in high demand in this
area because they often use intuition, instincts, and “non-linear” thinking strategies to
discover unknown classes of malware and unknown distribution patterns, which
existing autonomous malware discovery systems are not capable of matching. How-
ever, experts are busy and expensive. We need “one-shot” or even “zero-shot”
machine-learning algorithms to detect, map, and predict malware types and distribution
networks. Here we aim to develop an artificial intelligence virtual analyst that enables
a machine to learn cognitive discovery processes from human experts and then to map,
measure, attribute, and disrupt malware distribution networks.
A malware distribution network (MDN) is a set of top level domains (TLD) that have
been maliciously compromised. The TLD nodes are connected with directed edges where
the direction indicates the flow of malicious traffic across domains. These connected nodes
create a hidden structure used by malicious authors to spread malware globally. MDNs are
often used as a backend distribution network for attack campaigns such as botnets, spams,
and distributed denial of service attacks [1–4]. MDN is one of the key financial sources of
underground activities, malicious actors, and the overall hidden economy [3]. MDNs are
highly dynamic with domains and associated IP addresses changing constantly. Detection
of MDN is normally based on intensive manual analysis of log files, identifying the
malicious nodes, placing the appropriate edges, and plotting graphs. Usually, an MDN
won’t be discovered until the overall network distribution graph is plotted.
The data for this work is based on the Google Safe Browsing transparency report
(GSB) [5] along with malicious attribution provided by VirusTotal online service. This
work combines GSB and VirusTotal data to construct malware attributed MDNs over
the 9-month data collection period. The main contributions of this work include: (1) a
novel human-machine teaming model revealing existence of highly persistent sub-
MDN networks from the cloud-sourcing dataset, and (2) a combination of network
visualization, machine vision, and machine learning for dynamic data analytics.
2 Related Work
MDNs can be found in URLs [6, 7], botnets, pay-per-install, traffic redirection and
manipulation [8–10]. Provos et al. [11, 12] show that the existence of MDNs web-
based malicious campaigns, which led the launch of GSB service. In this study, we
enhance our previous work in visual analytics [21, 24] using a “breadcrumb-like”
pheromone model which traces topological changes over time.
In the wake of growing demand for human-machine teaming technologies for cyber
security, in April, 2018, DARPA announced the program CHESS (Computers and
Humans Exploring Software Security), which aims to develop human-machine teaming
system for rapidly discover all classes of vulnerability in complex software. In contrast
to the previous DARPA Cyber Grand Challenge (CGC) program that focuses on
autonomous systems, CHESS program includes humans in the discovery processes in
order to overcome the weaknesses of fully automated discovery systems.
MouselabWEB [8] is a process-tracing tool for extracting the process of decision
makers. The technology was originally developed for understanding the decisions made by
financial analysis, but has been used in other applications in use by consumers. Mouse-
labWEB has its disadvantages. For example, it is limited by predetermined structures and
orders of input as it was designed for business decision-making in general. Thus, it does not
have the flexibility for enabling the discovery process used by cyber analytic experts.
Perceiving Behavior of Cyber Malware with Human-Machine Teaming 99
Our human-machine teaming discovery system contains three components: (1) cyber
crawlers for data collection, (2) visual ontology for visualization and pattern extraction, and
(3) human-machine collaborative learning. Figure 1 illustrates the architecture of the system.
The cyber crawlers are designed to continuously collect data on malware patterns,
distribution, and attributions. The data is stored in JSON format. The dataset is con-
verted to texture images and graphs that reflect the changes over time. The visual
patterns are represented in visual ontology for human analysts to observe and for
machine vision to perceive automatically, including anomaly detection and clustering.
Finally, machine learning algorithms help human analysts to find similar dynamic
patterns in sub-graphs and computer the distances based on similarities.
In this study, we scripted crawlers to map malware attribution and malware distribution
network (MDN) data from available sources such as Google Safe Browsing (GSB) and
VirusTotal. The GSB reports contain timestamps, network nodes and links, while
VirusTotal reports contain malware type and submission timestamps. The crawler first
performs the data integrity checking by mapping abstract data into multi-layer textures
and detect anomalous events based on texture classification models such as energy,
contrast, and entropy. Then the crawler builds a malware distribution graph to show
which nodes are infected by which malware, their sources and destinations with
timestamps. This can help the analysts grasp a big picture of the malware distribution
and their evolutionary dynamics, including the sizes of the malware distribution
clusters, root nodes of malware host, super-nodes in the malware distribution network,
persistent links during the distribution, and the minimal-cut strategies to disrupt the
malware distribution network. Figure 2 illustrates the MDN generation process.
Fig. 2. Generating malware distribution network (MDN) from crawlers over time
Visual ontology in this study includes two layers of descriptions: First, at topological
level, e.g. the Top Level Domain Names and their graph connections; second, at the
binary code level, we will map the bits into texture and color patterns so that both
human analysts and computers can see it. Figure 3 shows visual ontology (in texture
and color) of the data from a mobile phone. After converting abstract data into visual
artifacts, we then use machine vision and learning algorithms to recognize data
structures, malware types, and topological dynamics Figs. 4 and 5.
In this study, we defined a set of visual objects as a language for describing patterns
and trends in the data. These visual objects include: (1) texture – mapping data to an
image with variation of intensity or color, (2) topology – graph relationship of a
connected network, (3) trace – duration of interaction between two nodes (entities), and
(4) dynamics – changes of state over time (e.g. texture, topology and trace).
We tested the texture pattern mapping with our collected data from Google Safe
Browsing (GSB). Each website entry in a captured collection has 24 base attributes
which GSB provides. We additionally calculate 2 attributes, root malicious and web
site type. Root malicious is a binary attribute specifying whether there are any inbound
edges for the website. Website types are determined by the label GSB assigns, which is
an attack site, intermediate site, or malicious node (default label when no other label is
assigned). Each non-quantitative value is valued either by assigning numerical category
value or by the attribute length. Then each value is normalized to be between 0 and 1
by dividing each attribute value by the maximum value across any collection. Each site
entry is then visualized in a 6 5 patch of 5 5 pixel squares. Each value is con-
verted to a brightness value for a white color and rendered into an image.
In the general data, the data quantity in the collection rarely fall below one quarter
of the largest collection we had across the months. The two other groups of raw data are
low quantity data. Some data collections are significantly less than others, but never-
theless cover a substantial group of websites. We believe these are anomalous data sets
given there is a sharp discontinuity in attributes. These corresponding collections may
have unique traits which are worth further investigation. Lastly we have the known
incomplete data sets, which may have occurred due to script crashes, computer outages
or any other external reason which generated near empty data sets. This visualization
helps us weed out those sets in an effective visual manner.
102 Y. Cai et al.
Fig. 5. Types of textural patterns of collections: normal data (left), anomalous data (middle) and
incomplete data – no data or very few entries (right)
Topology is an important visual language that can be used by humans and com-
puters. A graph is an abstraction of relationships between entities. Figure 6 shows
typical topological patterns of a graph. In this project, we continued our development
of a 3D graph visualization algorithm with pheromone-based trace model and dynamic
visualization for the GSB dataset of malware distribution networks.
In this study, we describe the MDN data structure as a list timestamp, source-
destination network node pairs with associate attributes. The computer reads the dataset
row by row to form the dynamic graph, where the distances between nodes are artic-
ulated with semantic attributes, for example, if the two nodes interact each other fre-
quently, they would move closer and closer until they reach a collision limit. We also
use the pheromone model [14] to visualize the persistency of malware on MDN edges
and nodes, where pheromone deposit represents how long the malware stays. The longer
the malware stays, the darker or bigger the shape. Pheromone decay represents that the
impact on the edge or node becomes lighter and smaller [15]. See Fig. 7.
Perceiving Behavior of Cyber Malware with Human-Machine Teaming 103
Fig. 7. The color map of the digital pheromone representation of persistency of the malware
distribution over the connected links (edges), ranging from 0 day to 20 days or more
The visualization of the malware helps the cyber analysts to discover the dynamics
of MDN clustering, for example, several large clusters grew in one to two weeks.
Smaller chains first formed over time before multiple chains gathered together to form a
large cluster. From our dataset in 2017, we observed a large centralized cluster of
similar domains in February, March, and June. This indicates potential large scale
malware attacks during those periods.
From visualization, we also observed the correlation between social events and
formation of large clusters of MDNs. For example, ten days after the US Presidential
Inauguration Day, January 20, 2017, there was a significantly large cluster of MDN that
continued till middle of February. See Fig. 8.
Fig. 8. A cluster of MDN on the US Presidential Inauguration Day, January 20 (left) and the
cluster formation on January 30 (right), 2017
In this study, we developed a novel machine learning algorithm that uses visual
ontology (e.g. texture and topology) as “feature vectors”, or hidden layers in a neural
network. It has more semantic knowledge, or augmented human intuition, than purely
numerical learning. Given samples of data, we can generate sequences of events in the
real-world environment. The input and output data can be used to train a machine
learning algorithm to recognize significant patterns. To build the model, we need to
extract features. For example, we can map the raw malware data to textured images and
104 Y. Cai et al.
X
255
variance ¼ ðpi mÞ2 N ðpi Þ
i¼1
X
255
flatness ¼ ðpi mÞ4 N ðpi Þ
i¼1
X
255
uniformity ¼ N 2 ð pi Þ
i¼1
X
255
entropy ¼ N ðpi Þ log2 N ðpi Þ
i¼1
We used the first order texture descriptors instead of the second order descriptors of
the grey-level co-occurrence matrix (GLCM) because the GLCM model is sensitive to
the arrangement of the neighborhood of the data. In our case, the order of the mapping
sequence is less important. From Fig. 9, we can easily distinguish anomalous data from
normal data.
Perceiving Behavior of Cyber Malware with Human-Machine Teaming 105
Fig. 9. Texture analysis of GSB dataset in entropy, variance, flatness, and uniformity
Here we consider two sequences of the topological attribute vectors, in this case,
the degree values: A ¼ a1 ; a2 ; a3 ; . . .; ai ; . . .an and B ¼ b1 ; b2 ; b3 ; . . .; bj ; . . .bm . We
start with the Dynamic Time Warping (DTW) algorithm to map the two sequences of
feature vectors on two sides of a two-dimensional grid. We then try to align them by
warping the time axes iteratively until the shortest distance between the two sequences
is found. Figure 11 shows topological changes (degree) over time in a sample malware
distribution network in 2017 and the DTW distances between the reference graph goo.
gl and the rest. As we can see, the bambaee.tk is the closest one. In this study, we
explore dynamic pattern recognition with multiple attributes and work on general
matching algorithms that can be applied to specific case studies in the following task.
The visual learning algorithm can be used for searching for similar dynamic patterns
and can also be used for self-organized clustering for exploring new visual ontological
classes or trends. See Fig. 11 and Table 1.
Fig. 11. DTW distance calculation indicates the closest two: goo.gl and bambaee.tk
7 Conclusions
Acknowledgement. The authors would like to thank research assistants Pedro Pimentel and
Sebastian Peryt for early prototypes. This project is in part funded by Cyber-Security University
Consortium of Northrop Grumman Corporation. The authors are grateful to the support from Drs.
Paul Conoval, Robert Pipe, and Donald Steiner. [DISTRIBUTION STATEMENT A] This
material has been approved for public release and unlimited distribution. Please see Copyright
notice for non-US Government use and distribution. Internal use: * Permission to reproduce this
material and to prepare derivative works from this material for internal use is granted, provided
the copyright and “No Warranty” statements are included with all reproductions and derivative
works. External use: * This material may be reproduced in its entirety, without modification, and
freely distributed in written or electronic form without requesting formal permission. Permission
is required for any other external and/or commercial use. Requests for permission should be
directed to the Software Engineering Institute at [email protected]. * These restrictions
do not apply to U.S. government entities. Carnegie Mellon® and CERT® are registered in the U.
S. Patent and Trademark Office by Carnegie Mellon University. DM19-0291.
Distribution Statement A: Approved for Public Release; Distribution is Unlimited; #19-0490;
Dated 04/17/19.
References
1. Gu, G., Perdisci, R., Zhang, J., Lee, W.: BotMiner: clustering analysis of network traffic for
protocol- and structure-independent botnet detection. In: Proceedings of the 17th USENIX
Security Symposium (Security 2008) (2008)
2. Gu, G., Zhang, J., Lee, W.: BotSniffer: detecting botnet command and control channels in
network traffic. In: Proceedings of the 15th Annual Network and Distributed System
Security Symposium (NDSS 2008) (2008)
3. McCoy, D., et al.: Pharmaleaks: understanding the business of online pharmaceutical affiliate
programs. In: Proceedings of the 21st USENIX Conference on Security Symposium, Ser.
Security 2012, pp. 1–1. USENIX Association, Berkeley (2012)
108 Y. Cai et al.
4. Karami, M., Damon, M.: Understanding the emerging threat of ddos-as-a-service. In:
Proceedings of the USENIX Workshop on Large-Scale Exploits and Emergent Threats
(2013)
5. Google safe browsing. https://developers.google.com/safe-browsing/
6. Peryt, S, Morales, J.A., Casey, W., Volkmann, A., Cai, Y.: Visualizing malware distribution
network. In: IEEE Conference on Visualization for Security, Baltimore, October 2016
7. https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=7870760
8. MouselabWEB. www.mouselabweb.org
9. Barzdinš, J., Barzdinaš, G., Cerans, K., Liepinš, R., Sprogis, A.: OWLGrEd: a UML style
graphical notation and editor for OWL 2. In: Proceedings of the 7th International Workshop
on OWL: Experiences and Directions (OWLED 2010), volume 614 of CEUR-WS (2010)
10. Kost, R.: VOM – Visual Ontology Modeler (2013). thematix.com/tools/vom/
11. Howse, J.: Visualizing ontologies: a case study. In: International Semantic Web Conference,
ISWC 2011, pp. 257–272. Springer, Berlin (2011)
12. van der Maaten, L.J.P., Hinton, G.E.: Visualizing non-metric similarities in multiple maps.
Mach. Learn. 87(1), 33–55 (2012)
13. Nataraj, L., et al.: Malware Images: Visualization and Automatic Classification, VizSec
2011, 20 July 2011
14. Cai, Y.: Pheromone-based visualization model of malware distribution networks. In:
International Conference on Computational Science, to appear on Springer LNCS (2018)
15. Wigglesworth, V.B.: Insect Hormones, pp. 134–141. WH Freeman and Company,
Holtzbrinck (1970)
16. TopQuadrant. TopBraid Composer. http://www.topquadrant.com/topbraid/
HackIT: A Human-in-the-Loop Simulation
Tool for Realistic Cyber Deception
Experiments
1 Introduction
Deception is an art of making someone believe in something that is not true, may
provide a promising real-time solution against cyber-attacks [1]. Deception involves
interaction between two parties, a target and a deceiver, in which the deceiver effec-
tively causes the target to believe in a false description of reality [1]. The objective is to
cause the target to work in such a way that is beneficial to the deceiver. Deception has
been used as an offensive and defensive tool in cyber world by hackers and defenders.
Hackers used deception for exploiting cyber infrastructure, stealing information,
making money and defaming people. The deception techniques used by hackers may
involve malware signature, conceal code and logic, encrypted exploits, spoofing,
phishing, and social engineering (e.g., by deceiving help desk employees to install
malicious code or obtain credentials). However, defenders used deception for securing
network infrastructure, luring hackers to fake information and understanding hacker’s
motives and strategies using tools such as honeypots. When used for defense, cyber
defenders may use feints and deceits to thwart hackers’ cognitive processes, delay
attack activities, and disrupt the breach process. When used for defense, deception may
be achieved through miss-directions, fake responses, and obfuscations [2]. These
techniques rely upon hacker’s trust in response from network, data, and applications
during an actual attack. To create deception for defense, security experts have been
using honeypots, fake servers that pretend to be real, for gathering intelligence about
hackers. Honeypots are one of the effective deception tools in the network defense to
lure hackers.
Decisions-making process of hackers and defenders in cyber world is complex task.
Understanding their decision process in such complex environment is challenging.
Simulation has been used as an effective way of understanding the hackers’ and
defenders’ decisions, testing new solutions for security, and training the models and
people in such complex task scenarios [3–10] have used behavioral game theoretic
approaches to understand the decisions of hackers and defenders in the abstract cyber-
security games. In such game theoretic approaches, the network structure and the set of
actions of hackers and defenders were abstracted to attack/not-attack and defend/not-
defend. Furthermore, the information provided to the participants was also abstracted.
The task simulated using game-theoretic approaches was less cognitively challenging
compared to the real cyber-security tasks. Thus, the conclusions made based on
behavioral game-theoretic approaches may or may not address the cybersecurity
problems. Another approach used to study hackers and defenders behaviour involve
complex real time tools such as NeSSi, Network Simulator-2/3, Cymulate etc. [11, 12].
However, using deception in uncontrolled environments makes it difficult to answer
cause-effect questions.
Aggarwal et al. [6] proposed HackIT tool to bridge the gap between behavioral
game-theoretic approaches and real-world cybersecurity tasks. HackIT tool provided
features to create more specific set of actions and information needed for cyber-security
tasks for both hackers and defenders. The HackIT tool was used to replicate the results
of a laboratory experiment using a deception game [3, 4]. Results revealed that the
average proportion of attacks was lower and not-attack actions were higher when
deception occurred late in the game rather than earlier; and when the amount or
deception was high compared to low [6]. This result found in an abstract simplified
scenario was replicated in a real-world simulation tool called the HackIT.
In [6], the HackIT tool was available for creating deception with limited number of
systems only and for single player games only. In this paper, we define the enhanced
capabilities of HackIT tool. Specifically, we detail how the HackIT tool is capable of
running experiments with different sized networks, different network configurations,
different deception strategies, and single player and multiplayer games. Thus, the
enhanced capabilities of HackIT tool can help us answer several questions such as the
effect of different network sizes and honeypot allocations on hacker’s performance and
the most effective way to present the “clues” of deception in network nodes.
In what follows, we first discuss the functioning and different features in the of the
HackIT tool. Next, we detail an experiment involving the HackIT tool to showcase its
HackIT: A Human-in-the-Loop Simulation Tool 111
capabilities. Furthermore, we detail the results from the experiment and discuss the
implications of our results for investigating the decision-making of hackers in the real
world.
2 HackIT
HackIT is a generic framework for cybersecurity tasks to study human learning and
decision-making of hackers and analysts. It represents a simplified framework con-
taining the most essential elements for creating cybersecurity scenarios: network nodes,
which represents the characteristics of real nodes; strategies, which can be configured
for creating deception; and commands, which are used for communication with the
network. The analyst’s goal in the HackIT task is to evaluate different deception
strategies and the hacker’s goal to identify the real network nodes and exploit them.
Hackers communicate with the network in HackIT using different commands and gain
information about different configurations. However, hackers are not aware of the
strategies used by analysts inside the HackIT scenario. Hackers basically learn about
these strategies overtime by playing different rounds. Thus, HackIT is a controllable
and flexible simulation tool with the capability of creating various network scenarios
and experiment with different techniques to lure hackers. Figure 1 shows the typical
flow of HackIT tool which uses the concept of stackelberg security games [13], where
first defenders create a network and use their defense techniques and next, hackers try
to exploit the security of the network.
Despite the simplicity of HackIT, the tool has the potential to simulate many real-
world dynamic situations in the laboratory: testing different proportion of honeypots in
the network, testing optimal placement of honeypots, different configurations of
honeypots, such as, easy to exploit ports on honeypots, and availability of fake files on
honeypots, availability of unpatched ports on honeypot etc. Specific characteristics of
HackIT tool are explained as followed:
Fig. 2. Different network sizes: (A) Small, (B) Medium, and (C) Large
However, in multi-player experiment setup, players are provided with a chat func-
tionality to share their knowledge as shown in Fig. 4. Hackers usually penetrate into the
network by organizing themselves into a group. Hackers in a group rely on information
gained from fellow hackers who have already penetrated into the network or are trying
to penetrate it.
2.6 Commands
The HackIT tool can run various network commands that include: nmap, use_exploit,
ls, and scp. Nmap is a network utility that shows the open ports, operating system type,
and services running on the specified webserver. The nmap utility also provides the list
of vulnerabilities on the corresponding webservers. The use_exploit command exploits
vulnerabilities of a system and helps attacker to gain access to a particular webserver.
Next, the ls command lists the files currently on the file system of the machine. The scp
command transfers files to the remote machine.
– Probing behavior: HackIT tool provides capability to record the probing behavior
of hacker participants. Probing is an important phase where hackers collect infor-
mation before launching a real attack. Using deception in probing responses and
analyzing hackers’ reaction towards deceptive responses is an important measure
provided in HackIT tool.
– Attack behavior: HackIT measures the attack behavior of hackers by recording
their attack actions, exploit information, their specific selection of targets or con-
figuration. HackIT tool also records the vulnerabilities exploited, and exploits used
by hackers. Defenders can analyze this data to study their attack patterns.
– Learning: HackIT tool can record overtime behavior of participants. Thus, it
provides data to analyze learning capabilities of hackers against different deception
strategies.
– Communication History: HackIT tool provides functionality to analyze the
chatting history of hacker to investigate that how hackers make decisions or for-
mulate strategies in team-based tasks.
– Time: HackIT tool also records time taken by hackers to exploit a system. Timing is
another measure to evaluate the success of deception.
Fig. 5. Experiment design using deception game with six rounds and two conditions, i.e. early
deception and late deception. D: deception present -: deception not present [1, 2].
First, the attacker probed the network using nmap command to gain information
about different webservers. Probing different webservers gave the information about the
operating system, open ports, services, and vulnerabilities. The information provided to
the attacker as a result of probing systems gave him an idea about the type of con-
figuration on the probed system. Once the attacker collects information about open
ports and services, he could attack a webserver by using the “use_exploit” command.
The use_exploit command exploited vulnerabilities on a system and helped the attacker
to gain access to that webserver. Next, the attacker could list different files on the
exploited webserver by using the “ls” command. Next, the attacker could transfer
required files containing credit card information (e.g., “pin.txt”) using the “scp”
command. After attackers copied the file from the exploited system, he was informed
whether he was successful or not in stealing a real credit-card file from the computer
via a text-based feedback.
3.3 Participants
Participation was voluntary and a total of 16 male participants participated in the study
that was advertised via an email advertisement. Out of the 16 people, 62.5% people had
taken a course in computer networks/security. The age of participants ranged from 18–
22 years. About 56% were 3rd year and 44% were final year undergraduate students
from Indian Institute of Technology Mandi, India. All the participants were remuner-
ated at a fixed rate of INR50 for their participation in the study.
3.4 Procedure
Participants were given instructions about their objective in the HackIT task, and they
were informed about their own action’s payoffs. Specifically, human hackers were
asked to maximize their payoff by stealing the real credit-card file from the network
over several rounds of play (participants were not aware of the endpoint of the game).
Each round had two stages: Probe stage and Attack stage. Hacker could probe the
network using “nmap” utility in first stage. After probing the webservers, he received
information about open ports, operating systems, services, and vulnerabilities associ-
ated with different webserver. Next, the hacker had to choose one webserver to exploit
and exploit webservers using “use_exploit” command during attack stage. Once the
webserver was exploited, hackers transferred the credit-card file to their remote
computer.
3.5 Results
Figure 6 shows the proportion of attacks on honeypot and regular webservers.
There was no difference in the proportion of attacks in late and early deception
conditions.
118 P. Aggarwal et al.
Timings of Deception
0.56 0.56
Proportions of Attack
0.60 0.44 0.44
0.40
Actions
0.20
Honeypot
0.00
Early Late Regular
Deception Deception
Conditions
Fig. 6. Proportion of attack actions on regular and honeypot webservers across rounds and
participants.
Next, we analyzed the exploits used during deception rounds and non-deception
rounds by participants. When regular (honeypot) systems are attacked during deception
rounds, that is called as deception failure (success). Figure 7a and b show the number
of regular attacks and honeypot attacks against different exploits in deception failure
and success, respectively. During deception failure, the remote_auth vulnerability was
more exploited in early condition compared to late condition and the pop3_version
vulnerability was exploited more in the late condition compared to early condition.
During deception success, the brute_force vulnerability was more exploited more in
early condition compared to late condition and the DOS_attack and sql_injection
vulnerabilities were exploited more in the late condition compared to early condition.
Figure 8 shows the number of attacks on regular webservers using different vul-
nerabilities. We found that during early deception conditions, mostly hackers used
remote_auth and drown_attack vulnerabilities. Furthermore, during late deception
condition, hackers used pop3_version and remote_auth vulnerabilities.
HackIT: A Human-in-the-Loop Simulation Tool 119
In this paper, we discussed HackIT, a HITL simulation tool with a potential to help
cyber-security researchers to investigate the decision-making of attackers and defenders
in real-world cyber-security scenarios. In this paper, we showed different features of
HackIT tool and different ways to conduct multiplayer experiments. We also showed a
concrete example of using HackIT to investigate the effects of timing of deception on
hacker’s decisions. We believe that HackIT tool would be helpful in creating other
cyber-security scenarios involving dynamic network sizes, dynamic network configu-
rations, and various deception strategies.
First, we simulated an information stealing scenario in HackIT. We found that the
attacks on regular and honeypots were no different in early and late deception condi-
tion. One likely reason for this result could be that participants perceived easy to attack
and difficult to attack vulnerabilities similarly. In fact, hacker participants exploited
remote_auth vulnerability to attack on regular machines and brute_force and sql_in-
jection vulnerabilities to attack honeypot systems. Furthermore, we found participants
attacked more number honeypot systems compared to regular systems.
Cybersecurity faces different open challenges while implementing the deception.
These challenges may involve the following questions: what an effective deception
strategy should be? when should the deception be used? how hackers can be deceived
during the probe phase? what are their probing patterns? and, how to make deception
less detectable? HackIT tool could provide a framework to investigate these questions.
One way to make deception less detectable is to have effective configuration and
content on deceptive nodes. HackIT tool could be used to identify effective configu-
rations and contents on honeypots to make them less detectable.
In future, we plan to perform a series of experiments involving participants per-
forming as attackers in other simulated network scenarios in HackIT. Here, we wish to
extend the HackIT tool to investigate the optimal proportion of honeypots and effec-
tiveness of deception in networks of different sizes. For example, a network could be
classified as small, medium, or large sized based on number of webservers present.
120 P. Aggarwal et al.
Acknowledgments. Research was partially sponsored by the Army Research Laboratory and
was accomplished under Cooperative Agreement Number W911NF-13-2-0045 (ARL Cyber
Security CRA). The views and conclusions contained in this document are those of the authors
and should not be interpreted as representing the official policies, either expressed or implied, of
the Army Research Laboratory or the U.S. Government. The U.S. Government is authorized to
reproduce and distribute reprints for Government purposes notwithstanding any copyright
notation here on. This research was also supported by the ICPS DST grant (T-533) from the
Indian Government to Dr. Varun Dutt.
References
1. Whaley, B.: Toward a general theory of deception. J. Strateg. Stud. 5(1), 178–192 (1982)
2. Rowe, N.C., Custy, E.J.: Deception in cyber attacks. In: Cyber Warfare and Cyber
Terrorism, pp. 91–96. IGI Global (2007)
3. Aggarwal, P., Gonzalez, C., Dutt, V.: Cyber-security: role of deception in cyber-attack
detection. In: Advances in Human Factors in Cybersecurity, pp. 85–96. Springer, Cham
(2016)
4. Aggarwal, P., Gonzalez, C., Dutt, V.: Looking from the hacker’s perspective: role of
deceptive strategies in cyber security. In: 2016 International Conference On Cyber
Situational Awareness, Data Analytics And Assessment (CyberSA), pp. 1–6. IEEE (2016)
5. Aggarwal, P., Gonzalez, C., Dutt, V.: Modeling the effects of amount and timing of
deception in simulated network scenarios. In: 2017 International Conference On Cyber
Situational Awareness, Data Analytics And Assessment (Cyber SA), pp. 1–7. IEEE (2017)
6. Aggarwal, P., Gonzalez, C., Dutt, V.: Hackit: a real-time simulation tool for studying real-
world cyber-attacks in the laboratory. In: Handbook of Computer Networks and Cyber
Security: Principles and Paradigms. CRC Press (in press)
7. Bye, R., Schmidt, S., Luther, K., Albayrak, S.: Application-level simulation for network
security. In: Proceedings of the 1st international conference on Simulation Tools and
Techniques for Communications, Networks and Systems & Workshops, p. 33. ICST (Institute
for Computer Sciences, Social-Informatics and Telecommunications Engineering) (2008)
HackIT: A Human-in-the-Loop Simulation Tool 121
8. Maqbool, Z., Makhijani, N., Pammi, V.C., Dutt, V.: Effects of motivation: rewarding
hackers for undetected attacks cause analysts to perform poorly. Hum. Factors 59(3),
420–431 (2017)
9. Dutt, V., Ahn, Y.S., Gonzalez, C.: Cyber situation awareness: modeling detection of cyber-
attacks with instance-based learning theory. Hum. Factors 55(3), 605–618 (2013)
10. Aggarwal, P., Maqbool, Z., Grover, A., Pammi, V. C., Singh, S., Dutt, V.: Cyber security: a
game-theoretic analysis of defender and attacker strategies in defacing-website games. In:
2015 International Conference on Cyber Situational Awareness, Data Analytics and
Assessment (CyberSA), pp. 1–8. IEEE. (2015)
11. Cymulate- Breach and Attack Simulation. https://cymulate.com. Accessed 15 Feb 2019
12. Issariyakul, T., Hossain, E.: Introduction to network simulator 2 (NS2). In: Introduction to
Network Simulator NS2, pp. 1–18. Springer, Boston (2009)
13. Tambe, M.: Security and Game Theory: Algorithms, Deployed Systems, Lessons Learned.
Cambridge University Press, Cambridge (2011)
14. Achleitner, S., La Porta, T.F., McDaniel, P., Sugrim, S., Krishnamurthy, S.V., Chadha, R.:
Deceiving network reconnaissance using SDN-based virtual topologies. IEEE Trans. Netw.
Serv. Manag. 14(4), 1098–1112 (2017)
Mathematical Model of Intrusion Detection
Based on Sequential Execution of Commands
Applying Pagerank
testing phase, with an accuracy rate greater than 90% and a false positive rate of
less than 4%. This shows that our model is effective and adaptable to the
dynamic behavior of the user. On the other hand, a variability in the execution of
user commands has been found to be quite high in periods of short time, but the
proposed algorithm tends to adapt quite optimally.
1 Introduction
Currently, several artificial intelligence techniques are applied, such as the k-nearest
neighbors (KNN), as detailed in their work [5]. Their proposal is a representation of
novel characteristics identifying the center of the data groups (cluster). This model
measures and adds two distances, where the first distance is a sample and its center are
the cluster of data. The second distance is determined between the information and its
nearest neighbor in the same cluster.
A function that defines each data of the cluster is generated and later is entered into
a KNN algorithm to obtain an intrusion detection with a quite optimal result. This
proposal obtains a fairly high precision and a very low false positive rate.
Another approach in the detection of intruders is the one published by [6], the same
that develops a new hybrid model that consists of the algorithms J48, Meta Pagging,
RandomTree, REPTree, AdaBoostM1, DecisionStump and NaiveBayes. The objective
of the study is to estimate the degree of intrusion scope in a computer network,
identifying specific characteristics for the study. The experimental results obtained in
this article showed that the proposed model minimizes computational complexity and
detection time. In addition, the detection rate is close to 99.81%.
A study on intrusion detection is the one carried out by [7], which proposes an
algorithm to detect anomalous behaviors of users in computer sessions. This model
allows to detect a behavior profile of each user and identifies small anomalies in
sequences of 2 commands within a user session. This model applies an algorithm based
on the probability of these sequences. The activities classified as possible anomalies are
verified by the application of Markov chains. This method has shown to be efficient in
terms of high detection accuracy and low false positive rate.
In this paper we propose to use a command execution database in a UNIX operating
system, where there are eight different users. With this information, a behavior profile
of the execution of each user is generated and the activities that are most executed are
determined. The execution probability of each command is calculated with the Page
Rank algorithm [1, 8], which has been the fundamental basis for this study.
The work is structured as follows: In Sect. 2 the methods and materials used in the
research are presented. The methodology used for the development of the intrusion
detection model using data mining and the Page Rank algorithm is described. Section 3
details the proposed mathematical model and raises the phase of training and testing.
Section 4 presents the results of the model with its accuracy and performance. In
Sect. 5, a discussion of the results is presented comparing with works already pub-
lished. Finally, we present the conclusions obtained in the research and future works.
The following section details the techniques used to analyze the information and
the algorithms that formed the basis of the development of the intrusion detection
model.
This section presents the database that was used to generate the intrusion detection
model. On the other hand, the techniques of artificial intelligence to perform data
mining are detailed, as well as the Page Rank algorithm which enables us to define the
profiles of each of the users.
Mathematical Model of Intrusion Detection 125
2.1 Materials
This section details the information that was used as a basis for the development of the
model, such as the training and testing of the proposal.
This database contains a set of eight users, which were collected from computers
with a UNIX operating system. This information belongs to the history commands
execution during a period of two years. This data set is sequential, that is, each
command was executed consecutively during the use of the computer. In addition, this
database has flags that identify the start of each of the Shell sessions of execution of
commands, using the tag ** SOF ** and ** EOF ** to determine the end of the
session. The sessions of each user are within a text file arranged by date [9]. An
example of the database is showed in Fig. 1.
**SOF**
cd
<1> # one "file name" argument
ls
-laF
|
more
cat
<3> # three "file" arguments
>
<1>
exit
**EOF**
The distribution of information from the Unix database is shown in Table 1, where
the length of the execution of commands in a session can be evidenced, grouped by
users. In addition, the number of commands used in the entire database can be observed.
The data structure with which we are going to work is shown in Fig. 2, where a
user session contains 1 or n number of commands executed. This describes a dynamic
behavior on the use of the operating system.
** SOF ** EOF
** f <1> f <1> f <1> **
** SOF ** EOF
** elm telnet <1> f <1> **
** SOF ** EOF
** cd <1> cd <1> ls cd **
** SOF teln ** EOF
** et <1> lo **
** SOF ** EOF
** f <1> date telnet <1> f **
** SOF ** EOF
** f <1> telnet <1> elm **
** SOF ** EOF
** elm vt100 elm telnet <1> telnet **
** SOF ** EOF
** f <1> date elm ls **
** SOF ** EOF
** elm f <1> date elm fg **
** SOF teln ** EOF
** elm f <1> date et <1> **
** SOF ** EOF
** lo **
** SOF ** EOF
** elm lo **
A user session is defined as Sum , where m is the number of sessions that contains the
history of each user u. Each session S contains 1 or n number of commands executed,
defined as Cn , where n is the position in which the command has been executed. A user
session would be defined as Sum ¼ fC1 ; C2 ; C3 ; . . .; Cn g. Within each user session Sum
there are several subsequences of two commands, and these subsequences are defined
as sb ¼ fCn ; Cn þ 1 g, which provide information on which command Cn þ 1 is executed
after the Cn command.
2.2 Methods
This section presents the data mining algorithms that are used to pre-process the
information in the UNIX database, as well as the PageRank algorithm which allows
defining the profiles of each user.
Instance and attribute selection algorithms
Multiple data mining techniques are necessary to perform the pre-processing of
data. These techniques allow us to identify which instances and attributes are correct
and provide a large amount of information to the study.
Mathematical Model of Intrusion Detection 127
0
X 0
R ðvÞ
R ðuÞ ¼ c v2Bu N
þ cEðuÞ ð1Þ
v
The following section presents the development of the proposed model using the
Unix database as a fundamental resource for the generation of user profiles and the
PageRank algorithm to identify the probability of execution of commands.
In this section we present the intrusion detection model by applying PageRank to the
database. As a first phase, a pre-processing of data must be performed in which the
instances with noise must be eliminated by applying the Automatic Noise Reduction
algorithm. The result of the elimination of 3% of data with noise in the information of
the 8 users has allowed a more optimal database.
The Greedy Stepwise algorithm has been applied to identify the most important
commands, of which 17 commands has been identified as the most executed by the
128 C. Guevara et al.
users. With this database, 43% of the information that was not relevant to the study has
been eliminated.
To identify the probability of the most used commands by each user, it is necessary
to make a square matrix Mn;n , which contains 17 columns and 17 rows. The matrix Mn;n
will be described by each command Cn , where if a command Cn þ 1 is executed after a
command Cn in the matrix Mn;n þ 1 it will be assigned the value of 1, otherwise the
value of 0. An example is presented in the following matrix, where we only work with
5 commands (n ¼ 5):
2 3
0 0 1 1 1
60 0 0 0 17
6 7
M5;5 ¼6
61 0 0 1 177
41 0 1 0 05
1 1 0 0 0
Afterwards,
P an addition will be made per column of the matrix M, defined as
sumi ¼ ni¼1 Mn;i , and then all the values of each column will be divided by sumi ,
obtaining the following matrix M 0 :
2 3
0 0 1=2 1=2 1=3
6 0 0 0 0 1=3 7
6 7
M0 ¼ 6
6 1=3 0 0 1=2 1=3 7
7
4 1=3 0 1=2 0 0 5
1=3 1 0 0 0
The next step is to calculate the v value of the pagerank, as shown in Eq. (1). We
must take this matrix M 0 to a system of auto vectors defined as M 0 v ¼ kId v, where
k 2 R. We add up the additive inverse to the equation obtaining a homogeneous system
and calculating the value of pagerank v, as shown:
ðM 0 kId Þv ¼ 0
02 3 2 3 10 1
0 0 1
2
1
2
1
3 k 0 0 0 0 C1
B6 0 0 0 0 7 60
1
k 0 0 07 CB C 2 C
B6 1 7 6
3 7 CB C
B6 7 60 k 07 CB C
7 CB C 3 C ¼ 0
1 1
B6 31 0 0 2 7 6
3 0 0
@4 0 1
0 05 40 0 0 k 5
0 A@ C 4 A
3 2
1
3 1 0 0 0 0 0 0 0 k C5
Where the value v of each of the commands Cn is calculated, obtaining the fol-
lowing result C1 ¼ 6, C2 ¼ 1, C3 ¼ 5:33, C4 ¼ 4:66 and C5 ¼ 3. Obtaining the
probability of each command PðC1 Þ ¼ 1, PðC2 Þ ¼ 0:16, PðC3 Þ ¼ 0:88, PðC4 Þ ¼ 0:77
and PðC5 Þ ¼ 0:50. With these results, we can see which command is the most and least
used by the user.
For the Unix database, this process has been carried out with the eight users, where
the most commonly used commands by each user have been detected. To identify the
normal behaviors, it has been proposed to verify if the commands per each user have a
Mathematical Model of Intrusion Detection 129
probability of being executed in new sessions. In the event that the probability is low or
non-existent, a warning message of “possible intruder” will be launched.
The experimental results of the intrusion detection model are presented in the next
section. In addition, the results obtained with other previously published works will be
discussed.
4 Results
The results obtained in this study have been quite optimal, the tasks most executed by
the user have been identified in a more precise way. In this way an anomalous and
possibly intrusive behavior can be identified. Table 2 shows the results of detection
accuracy in both algorithm training and testing.
Table 2. Results of the training and tests of the algorithm proposed with the UNIX database.
Training Test
User Accuracy Error Accuracy Error
USER0 89,71 10,29 89,00 11,00
USER1 88,56 11,44 90,21 9,79
USER2 88,89 11,11 88,12 11,88
USER3 90,01 9,99 90,45 9,55
USER4 89,46 10,54 89,31 10,69
USER5 88,00 12,00 88,02 11,98
USER6 87,88 12,12 87,32 12,68
USER7 90,12 9,88 90,05 9,95
Average 89,08 10,92 89,06 10,94
With these results it has been shown that the accuracy rate of this algorithm is
optimal with an average of 89.08% in training and 89.06% in the testing phase,
compared with the results obtained by [7], of correctly classified of 94.78%.
The conclusions obtained in the development of this research are presented in the
following section, as well as the future lines of research that emerged during the study.
In this work we have identified points of great importance to take into consideration,
because they provide much information to efficiently detect an intrusive behavior. As a
fundamental basis, the execution of tasks in a sequential manner is essential to develop
a profile of a user’s behavior in a computer system, due to in small sequences the
human being repeats activities constantly during the use in a system. This study
showed that there are commands that are the most and least used by each user, and also
vary significantly depending on the position and responsibilities. The application of the
pagerank algorithm allows us to design a dynamic profile that evolves over time.
130 C. Guevara et al.
This is a great advantage, since a user is changing constantly their way of inter-
acting and working. The results obtained are optimal, but they are still a theoretical
approach to design a complete and efficient user profile. As future research lines, we
intend to apply hidden markov networks to accurately determine an intrusive behavior.
In addition, multiple generic user profiles will be generated according to position and
responsibilities, which will allow a possible intrusion to be detected in a more real way.
References
1. Gleich, D.F.: PageRank beyond the web. SIAM Rev. 57(3), 321–363 (2015)
2. Ashfaq, R.A.R., Wang, X.-Z., Huang, J.Z., Abbas, H., He, Y.-L.: Fuzziness based semi-
supervised learning approach for intrusion detection system. Inf. Sci. (Ny) 378, 484–497
(2017)
3. Kim, G., Lee, S., Kim, S.: A novel hybrid intrusion detection method integrating anomaly
detection with misuse detection. Expert Syst. Appl. 41(4), 1690–1700 (2014)
4. Elhag, S., Fernández, A., Bawakid, A., Alshomrani, S., Herrera, F.: On the combination of
genetic fuzzy systems and pairwise learning for improving detection rates on intrusion
detection systems. Expert Syst. Appl. 42(1), 193–202 (2015)
5. Lin, W.-C., Ke, S.-W., Tsai, C.-F.: CANN: an intrusion detection system based on
combining cluster centers and nearest neighbors. Knowl.-Based Syst. 78, 13–21 (2015)
6. Aljawarneh, S., Aldwairi, M., Yassein, M.B.: Anomaly-based intrusion detection system
through feature selection analysis and building hybrid efficient model. J. Comput. Sci. 25,
152–160 (2018)
7. Guevara, C., Santos, M., López, V.: Data leakage detection algorithm based on task
sequences and probabilities. Knowl.-Based Syst. 120, 236–246 (2017)
8. Page, L., Brin, S., Motwani, R., Winograd, T.: The PageRank citation ranking: bringing
order to the web, November 1999
9. Aeberhard, S., Coomans, D., Vel, D.: UCI Machine Learning Repository: UNIX User Data
Data Set. https://archive.ics.uci.edu/ml/datasets/UNIX+User+Data. Accessed 17 Dec 2018
10. Zarkami, R., Moradi, M., Pasvisheh, R.S., Bani, A., Abbasi, K.: Input variable selection with
greedy stepwise search algorithm for analysing the probability of fish occurrence: a case
study for Alburnoides mossulensis in the Gamasiab River, Iran. Ecol. Eng. 118, 104–110
(2018)
11. Xinchuan, Z., Martinez, T.: A noise filtering method using neural networks. In: IEEE
International Workshop on Soft Computing Techniques in Instrumentation, Measurement
and Related Applications, SCIMA 2003, pp. 26–31 (2003)
Investigation and User’s Web Search Skill
Evaluation for Eye and Mouse Movement
in Phishing of Short Message
Abstract. There are many studies on eye and mouse movement. However,
there are not many studies that try to evaluate the skill of Web search while
considering the relationship between the line of sight and the movement of the
mouse. In this study, we analyze the data acquired from the viewpoint of
the differences in information literacy of subjects and investigate the method of
quantitatively evaluating the skill of web search.
Keywords: Eye and mouse movement data Web search skill Skill evaluation
1 Introduction
With the development of ICT technology and the spread of smart devices, various
kinds of information are accumulated on the Internet. Big data analysis is very
important technology of finding valuable information from such a large amount of data,
and research is continuing in various fields. Although development of such big data
analysis technology can be said to be an important social theme, it can be said that the
web search capability that enables users to quickly find the necessary information from
the Internet is also an important skill. Ease of finding information on Web pages can be
quantitatively evaluated by Fitz’s law [1]. Fitz’s law can also be used to improve web
pages to make it easier to find information through web design. However, the infor-
mation required by individual users may be included in web pages that do not consider
UI. Also, because there are some information that are hard to get caught in the top
ranking in web search, skill related to web search is important in order to obtain
appropriate information quickly.
The purpose of this study is to investigate how user’s web search skills can be
reflected in data by collecting gaze and mouse data during web search.
It is well known that the gaze and mouse movement data has correlation [2]. This
study had investigated the process of web search using the theme that is difficult to hit
by web search. These days, the phishing method using SMS is becoming a social
problem. It is not easy to find the evidence of SMS phishing by web search. So, we had
treated the theme of information search on phishing using SMS.
Some SMS phishing is observed when an emergency such as an earthquake or a
heavy rain. Because the amount of information sent by SMS is small, at first glance
there is important information on emergency situations and the link URL. Since SMS
can write only a small amount of information, it can be said that it is a natural way to
have additional information refer to URL of link. Users who are familiar with phishing
techniques will check to see if other users are providing information on the Internet or
not by without clicking on that URL. So, by investigating the process of how to
perform web searching, we will study what kind of difference can be seen between
people with high IT literacy and people with low IT literacy.
2 Related Study
In this research, we had collected the data on gaze and mouse movement during web
searching, and investigate the characteristics of those data. Conventional research on
gaze and mouse movement and conventional research on Web search will be intro-
duced in this section.
3 Experiment
We did 2 types of experiments as the below in order to collect the data of eye and
mouse movement when a user searches information on the web.
Experiment 1
Each subject was shown the screenshot of SMS whose content is phishing, and we
examined that whether he/she found the screenshot the phishing message or not
(Fig. 1).
Experiment 2
After the Experiment 1, each subjects searched the evidence indicating that the
SMS is a phishing message. All subjects used Google or Bing as the search engine.
[Environment]
Eye-tracker: Tobii X2-30 Compact Mouse Cursor Tracing: UWSC OS: Windows10
Pro 64bit Number of Subjects: 4
Fig. 2. AoI on the phishing SMS. The source address is written in the blue area, the received time
is written in the orange area and the contents are written in the red, yellow sky blue and green area.
All subjects had focused attention on the word “Emergency Earthquake Warning”,
in common. It can be seen that subjects with high literacy gaze at the domain part of the
e-mail address. Because of the small amount of data, there are no major features of AoI
other than gaze to the domain. However, it can be said that the result that we cannot
deny the possibility of evaluating the presence or absence of literacy from the place the
subjects gaze eye movement is obtained.
What is common to all subjects is that the mouse movement distance has little effect
on literacy differences. And, it can be said that the feature of subjects with low literacy,
the eye movement distance is small. Therefore, subjects with low literacy have no
correlation between eye and mouse movements. It can be said that the characteristic of
presence or absence of literacy like this has regression line equation.
5 Conclusion
In this study, we had analyzed user’s eye and mouse movement data during web search
on SMS phishing, and examined its features. However, since the number of subjects is
small, it is the our future work to investigate whether the same result can be obtained
even if the number of subjects is increased.
References
1. Fitts, P.M.: The information capacity of the human motor system in controlling the amplitude
of movement. J. Exp. Psychol. 74, 381–391 (1954)
2. Chen, M.C., Anderson, J.R., Sohn, M.H.: What can a mouse cursor tell us more?: correlation
of eye/mouse movements on web browsing. In: Proceedings of CHI 2001, pp. 281–282. ACM
(2001)
3. https://sitest.jp
4. https://www.microsoft.com/ja-jp/enable/products/windows10/eye.aspx
5. Huang, J., White, R., Buscher, G.: User see, user point: gaze and cursor alignment in web
search. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems,
pp. 1341–1350 (2012)
6. Oki, M., Takeuchi, K., Uematsu, Y., Ueda, N.: Mobile network failure detection and
forecasting with multiple user bahavior. In: Proceedings of the Annual Conference of JSAI,
JSAI2018(0), 1D105 (2018)
7. Web Search Skill Evaluation from Eye and Mouse Momentum. In: Proceedings of the 2nd
International Conference on Intelligent Human Systems Integration (IHSI 2019), pp. 884–889
(2019)
Author Index
A I
Agarwal, Vaibhav, 109 Ikeda, Katsumi, 131
Aggarwal, Palvi, 109 Ishikawa, Hiroki, 131
Aguilar-Galvez, Fernando, 122
Arias, Hugo, 122 K
Kanahama, Nobuhiro, 131
C Katoh, Daiki, 131
Cai, Yang, 97
Cain, Ashley A., 79 L
Caporusso, Nicholas, 48, 87 Lehmann, Michal, 38
Casey, William, 97 Leukfeldt, Rutger, 16
Chalco-Torres, Lorena, 122
Corradini, Isabella, 59 M
Mancuso, Vincent, 69
D Matsuda, Takeshi, 131
Dutt, Varun, 109 McGuire, Sarah, 69
Morales, Jose A., 97
Morgan, Phillip L., 3
E
Eto, Masashi, 131
N
Ezer, Neta, 97
Nardelli, Enrico, 59
F O
Fandakly, Thaier, 48 Ortiz, Dioselina Pimbosa, 122
P
G
Patterson, Wayne, 28
Gautam, Aksh, 109
Gonzalez, Cleotilde, 109
R
Guevara, Cesar, 122
Ramirez-Morales, Ivan, 122
H S
Hanada, Tomohiro, 131 Satoh, Hironobu, 131
Hidalgo, Jairo, 122 Sonoda, Michio, 131
Soteriou, Robinson, 3 W
Staheli, Diane, 69 Wang, Sihan, 97
Stainbrook, Michael, 87 Williams, Craig, 3
Still, Jeremiah D., 79 Winston-Proctor, Cynthia E., 28
Szaniawska, Katarzyna, 38 Wisniewska, Malgorzata, 38
Wisniewski, Zbigniew, 38
U Y
Ushigome, Ryutaro, 131 Yandún, Marco, 122
Z
V Zapata-Saavedra, Lorena, 122
van der Kleij, Rick, 16 Zhang, Qiyuan, 3