Panic Passwords: Authenticating under Duress
Jeremy Clark*
School of Computer Science
University of Waterloo
j5clark@cs.uwaterloo.ca
Urs Hengartner*
School of Computer Science
University of Waterloo
uhengart@cs.uwaterloo.ca
Panic passwords allow a user to signal duress during authentication. We show that the well-known model of giving a user two passwords, a `regular' and a `panic' password, is susceptible to iteration and forced-randomization attacks, and is secure only within a very narrow threat model. We expand this threat model significantly, making explicit assumptions and tracking four parameters. We also introduce several new panic password systems to address new categories of scenarios.
As important services and sensitive data congregate online, attackers have an increasing incentive to obtain the passwords that protect these services and data. Panic passwords are a mechanism to allow a user to use a special type of password to signal to the server that her password is being entered as the product of a coercive action. While panic passwords are currently used in home security systems to trip a silent alarm, they could also find application online. For example, a shortcoming of Internet-based voting is the move from a private voting booth to an open environment where voters could be coerced or bribed to demonstrably vote a certain way. A panic password could allow a voter to cast a coerced ballot as if it were her real vote, while in reality the ballot is spoiled by the server.
This paper is motivated by a deficit of academic literature exploring the topic of panic passwords, the underlying threat models panic passwords address, and possible schemes to mitigate the threat while being usable. In particular, it is perturbing that the best-known example of a panic password scheme, one where the authenticator knows two passwords: a `regular' one and a `panic' one, is very easily defeated by coercing the victim to authenticate twice using different passwords each time.
Contributions: This paper makes the following new contributions to the academic literature on panic passwords:
• a thorough threat model for categorizing scenarios of coercion and undue influence,
• the application of an iteration and forced-randomization principle to panic passwords,
• the precise criteria for using the known scheme 2P,
• the introduction of three new panic password schemes: 2P-lock, 5-dictionary, and 5-click.
To make our contributions concrete, we will consider several representative scenarios where panic passwords could be employed. Our ultimate interest, however, is in Internet-based voting, which has been widely used in Estonia and in a limited capacity in Australia, Switzerland, and Canada. We will examine the potential of panic passwords to mitigate the improper influence problem: that removing the privacy of the voting booth opens the voting process up to the possibility of voter coercion and vote buying schemes.
Panic passwords are alternatively referred to as distress passwords or duress codes in the academic, commercial, and military literature. Due to their applicability in military and intelligence scenarios, it is not possible to determine the exact extent to which panic password schemes have been studied, as any such work would be classified. However a survey of the non-classified literature reveals very little.
Panic password schemes show up in patented technology, usually as one feature of a larger system. The exact method is often not included. A selection of patents include: ATMs which display a list of words from which the user may choose a predefined one for normal authentication and any other word to signal duress and to limit the amount of cash available for the transaction [10]; a home security system that incorporates a basic panic password scheme [7]; the use of panic passwords to authenticate over a network where panic mode relays the user to a different database than the normal one [6]; the use of a panic password to keep data protected on a mobile device [8]. Chaum mentions duress codes in passing for preventing coerced transactions using a credential [4]. In a policy recommendation, the Foundation for Information Policy Research suggest that files hidden on a hard-disk (using a steganographic file system [1]) could be erased upon entering a panic password [11].
Each of the scenarios considered herein involves three participants: Alice, Bob, and Oscar. Alice is typically a human and wishes to communicate, typically for the purpose of authentication prior to gaining access to a resource, to Bob, who may be human, a server, or a software application. Oscar is an opponent in the system and he attempts to coerce Alice's communications with Bob in order to benefit his nefarious purposes. We will also consider less coercive scenarios where Alice wishes to sell her access rights to Oscar. We assume that Bob is a trusted entity and, in particular, is not in collusion with Oscar.
Consider the password space , where each
is in one of
two sets,
for valid or
for invalid. As in any password scheme, a regular password
is selected
from
and is
included in
. Entering an
invalid password results in an error (observable to Alice and Oscar) and has no
further consequence. To extend the notion of passwords to panic passwords, we
define the following concepts.
Definition 1 Panic Password: a covert communication from Alice to Bob over an observed channel
indicating an abnormal state. We define a panic password as and include
it in
. A scheme
may include more than one panic password. All elements of
other than
and the
panic password(s) are invalid.
Definition 2 Unobserved Reaction: in reaction to receiving a from Alice,
Bob may engage in an unobserved reaction
, where
is the set of all such reactions and
is the
regular (non-panic) reaction. This reaction is unobservable to Alice or Oscar.
A scheme will include a single regular reaction and any number of panic
reactions (i.e., none, one, or multiple reactions).
Definition 3 Observable Response: in reaction to receiving from Alice,
Bob may respond to Alice with a panic response
, where
is the set of all responses and
is the
regular response. This response is observable to Alice and Oscar in principle,
however it is not necessarily distinguishable to Oscar which response is which.
A scheme will include a single regular response and any number of panic
responses.
Consider again Internet-based voting. To
authenticate to the voting service (Bob), assume Alice has a set of two
passwords: a normal password and a single panic password, . Entering
will cause
Alice's vote to be cast correctly, while entering
could trigger
an unobserved reaction from Bob,
: for example,
Bob could mark the ballot as spoiled or could alert the authorities. Entering
could also
trigger an observable response
: Bob may
falsely inform Alice that her vote has already been cast and she may not modify
it, or he may report that the server is down and unable to accept votes at this
time. In other areas where panic passwords are used, an unobserved reaction
could be a silent alarm, while an observable response may be modified access
control permissions or an upper limit on funds available for a financial
transaction.
A typical transaction will take the form,
(1)
(2)
(3)
Our threat model is based on four pragmatic assumptions.
Kerckhoffs' principle: The details of the authentication system that is in place is public information and known to all participants, including Oscar. Alice and Bob retain a shared secret (e.g., a password or set of passwords) that is both private and the foundation upon which the security of the system rests.
Observational principle: Alice communicates with Bob over a semi-private channel that can be observed by Oscar prior to the password being secured (i.e., encrypted or hashed). Namely, Oscar can observe the password(s) used by Alice and could even enter in Alice's password himself, after coercing her into revealing it.
Iteration principle: Unless explicitly prevented by the underlying system, Oscar is not bound to a single instance of coercion against Alice. He may force Alice to authenticate multiple times. Combined with the observational principle (Assumption 2), Oscar can force Alice to use a different password each time.
Forced-randomization principle: Oscar can choose to eliminate any strategy Alice may employ through the order in which she reveals the passwords she knows. For example, Oscar may force Alice into writing down a set of passwords so that he can randomly choose the order in which to iterate through them. The assumption is that this option is available to Oscar, not that he will necessarily employ it.
Parameter 1 |
Parameter 2 |
persistent |
|
non-persistent |
|
|
|
Parameter 3 |
Parameter 4 |
|
screening |
|
signaling |
|
|
|
Table 1: Summary of parameters to the threat model.
The underlying threat model is dependent on the scenario where a panic password system is employed. We identify four parameters along which the threat model could vary. They are summarized in Table 1 .
Oscar's Persistence: We can
define the period of time under which Alice is expected to be coerced by Oscar
as beginning at . In some
scenarios, Oscar may be non-persistent and limited in time to
, as in the
example of Oscar coercing Alice into withdrawing funds from an ATM. In other,
typically less-threatening, scenarios, Oscar may be persistent in his
coercion for an arbitrarily long period of time, as in the example of a
employer influencing the vote of her employee. A panic password scheme that
disables access to an account could deter a non-persistent attacker if the
account were disabled for a period of time greater than
but this
measure would not effectively deter a persistent attacker.
Bob's Reaction and Response: We
can summarize Equations 1 to 3 by defining a transaction as a tuple of parameters,
such as to mean Alice
submits non-panic communication and Bob reacts with the regular reaction and
response. If Alice presents a panic password, Bob has three options. He could
perform an unobserved reaction but not change his observable response,
; he could
modify his response but not commit any unobserved reaction,
; or he could
do both,
.
Oscar's Goal: Oscar's goal,
denoted , in coercing
Alice may take different forms. His proximate goal is for Alice to enter
but since he
reserves the option of iteration, we must distinguish between a few scenarios.
These distinctions are dependent on his ultimate goal with respect to Parameter
2.
If Bob performs a panic reaction to a
panic password, Oscar could have one of two goals: to prevent this unobserved
reaction from ever occurring in the set of transactions ,
; or to achieve
at least one regular reaction,
. If Oscar is
limited to a single transaction,
, these goals
are equivalent. However under the Iteration Principle, Oscar may force Alice to
authenticate more than once and these goals are no longer equivalent when
. If
were a silent
alarm, Oscar would have the first goal of preventing any occurrence of
. However if
was the
reaction of spoiling a ballot in an Internet-based election, Oscar may have the
second goal: he could force Alice to re-authenticate many times with different
passwords and vote each time, hoping that eventually she enters
and this
transaction will be counted as a valid vote.
If Bob responds differently to a panic
password than to a regular password in parameter 2, Oscar also has one of two
goals: to prevent any observable panic response or to achieve
at least one regular response
. As an example
of the first, consider the case where Alice has data hidden in an encrypted
partition on her harddrive. The regular response could reencrypt this data and
then provide access to it, while a panic response could overwrite the most
sensitive data with random data and provide access to only the less sensitive
data. Oscar's goal would be to prevent a panic response, as he cannot recover
from such a state. Alternatively, a panic response could be less permissive
access control permissions than is regular, and Oscar's goal will be to apply
iteration until he gets full access.
Screening vs. Signaling: A proper panic password scheme should cause Alice's use of a panic password to be indistinguishable by Oscar from use of her valid password. If Oscar coerces Alice into using her regular password, he succeeds by being able to distinguish (or convince Alice he can distinguish) the use of a panic password. Alice succeeds by not having to enter her regular password. Any mechanism that allows Oscar to separate panic and regular passwords is called a screen. Alternatively, Alice may want to prove to Oscar she is not using a panic password. Here Alice's goal is to demonstrate to Oscar that she is entering a regular password, and any mechanism she can use to accomplish this is called a signal. In Internet-based voting, voter-coercion is based on screening, whereas vote-selling is based on signaling. Although there is no duress in vote-selling schemes, the availability of panic passwords to prevent voter-coercion offers the additional feature that Oscar cannot determine if he is actually buying a legitimate vote from Alice, who could cheat him with her panic password. Thus, signal prevention is often a free additional property of panic password schemes.
We now consider several illustrative scenarios where panic passwords can be applied. These scenarios were chosen to represent a class of specific threat models to which a particular panic password scheme applies.
Consider a scenario with panic
communication such that
. Our approach
to this problem is invariant with respect to the persistence of Oscar or the
prevention of signals as opposed to screens. The approach also applies in an
equivalent manner to the scenarios:
such that
and
such that
.
Working Example: Consider the scenario where Alice is an employee at an office
building with a security alarm that must be deactivated with a numerical
password (entrance code) upon entry. Alice wants to communicate a forced entry
which alerts the alarm company or law enforcement () but still
grants Oscar access to the building (
). Oscar's goal
is to prevent the silent alarm as he cannot escape on time to avoid detention.[1]
Security alarms on the market already handle this through a system we will call 2P for two passwords.
2P: Alice knows two passwords . For normal
authentication purposes, Alice uses
. To
communicate panic, she uses
.
In the working example, Alice could use
a four digit PIN, , as
and invert the
order of the last two digits to construct
:
. Under our
first security assumption, Kerckhoffs' principle, Oscar is assumed to know the
structure of the system while not knowing
or
. From the
observational principle, Oscar can observe Alice enter a password,
, but cannot
determine if it is
or
. If access is
granted, Oscar can learn that it was one of the two passwords. Since Oscar's
goal is to prevent the silent alarm,
, forcing Alice
to enter a second password after entering the first (assumption 3) will only
assure he does not achieve his objective. Under these three security
assumptions, the 2P scheme succeeds at mitigating the threat.
The fourth assumption, however, presents
a problem. Since Oscar knows the scheme being used, he can ask Alice to reveal
the two passwords she knows. He cannot distinguish them but instead of him
allowing Alice to enter the password of her choosing, which presumably would be
, he can
randomly choose one. This allows him to achieve his goal half of the time on
average. However this probability could be reduced by having a larger set of
panic passwords, as will be examined below, and the significant probability of
being caught that 2P provides is likely a large enough deterrent for scenarios
in this class.
Consider a scenario where Oscar has a
non-persistent influence and his goal is either or
. This scenario
differs from the previous one since Oscar is also not attempting to prevent a
or
, only to gain
a clean
or
as the
situation dictates (thus, the outcome in footnote 1 would be acceptable).
Working Example: Consider the scenario where Alice is forced to withdraw money from an ATM by Oscar. In addition to wanting to signal a silent alarm (which we assume Oscar is not trying to prevent, as he believes he can escape on time), the ATM issues marked bills in place of real money. Oscar has a method for distinguishing marked bills but cannot do it on the spot. Therefore, his goal is only to escape with unmarked money even if he is not initially sure which bills are which.
The 2P system is inadequate in this
scenario as Oscar can conduct an attack using the iteration principle
(assumption 3). He will first force Alice to type in her PIN and withdraw
money, and then knowing the 2P system is in place (assumption 1), he will force
her to authenticate a second time while observing that she enters a different
PIN (assumption 2). Alice will have no choice but to enter on one of the
two occasions and Oscar will achieve his goal.
To address this threat, we can exploit
the fact that Oscar is non-persistent and must end his coercion at some point, , to escape.
Specifically if Alice's account is temporarily locked until some time
, Oscar will
not achieve his goal. We must, however, exercise caution in how locking of an
account is triggered. A naïve approach may be to lock the account after
receiving
, however Oscar
could use this information to screen
from
. He could
demand that Alice enters a password that does not produce a locked account by
threatening retaliation against her if her password does, thereby accomplishing
his goal of getting unmarked money. For this reason, the event that locks the
account must be invariant to the type of the passwords being entered.
2P-lock. Alice knows two passwords . For normal
authentication purposes, Alice uses
. To
communicate panic, she uses
which causes
or
as the
scenario dictates (but does not not lock the account). If Alice authenticates
multiple times within a window of time,
, using the
same password each time, her account will not be locked. However upon using two
different passwords within
, the account
will be made inaccessible for a period of
such that
.
The 2P-lock scheme is designed
specifically to thwart an iteration attack. Once Alice has entered one
password, forcing her to enter the other password is futile as it will only
lock the account. Furthermore, the locking will occur regardless of whether
Alice initially entered or
---making it
impossible for Oscar to determine if he achieved his goal by the behaviour of
the scheme. Oscar can conduct a forced-randomization attack, giving him
probability 1/2 of receiving unmarked money, however the scheme could include
more than two panic passwords which would lower this probability. Also, Oscar
could coerce Alice when he knows she has had recent (less than
ago)
communications with Bob. Since Oscar knows Alice has recently entered
, he knows that
any password that locks the account is a panic password, giving him the same
screening and coercion abilities that he had under 2P. However if Oscar could
observe when Alice communicates with Bob, he could more simply hijack her
account after she has authenticated and before she logs out.
Consider a scenario with panic
communication and a
persistent adversary with goal
. In addition,
preventing both signals and screens is important.
Working Example: Consider the previously defined improper influence problem in Internet-based voting. Alice could be coerced by Oscar into voting a certain way. Alternatively, Alice may want to sell her vote to Oscar by casting her ballot in his presence. Bob's observable response will be the report of a successful casting of Alice's vote, but he will take the unobserved reaction of disregarding any votes cast under a panic password.
In this case, we assume Oscar is
persistent, meaning he can observe Alice for extended periods of time. In the
working example, he could be her employer, an intrusive partner, or a union
representative. For this reason, locking the account until is not
guaranteed to be effective. It is also not prudent as it would easily allow an
adversary to prevent Alice from voting by coercing her into repeatedly locking
herself out of her account---an effective denial of service attack. Therefore
2P-lock is not suitable. The nature of the system does not preclude the
iteration attack either, rendering 2P ineffective as well.
The problem with a persistent attacker
is that he could eventually exhaust Alice's memory of panic passwords, forcing
her to enter eventually.
Therefore a suitable scheme for this model would equip Alice with an
arbitrarily large number of panic passwords, so that she could produce a
virtually endless list of them.
P-Compliment. Alice knows one password, , which she
uses for normal authentication purposes. To communicate panic, she uses any
other password:
. There are no
invalid passwords in this scheme.
This system allows Alice to enter any
password other than her real password to communicate a panic state. The
drawback of P-Compliment is that if Alice mistakenly enters the wrong password,
by definition of the problem, she cannot distinguish the panic state from the
normal one. From a usability perspective this is undesirable. In the working
example, given that there is no consequence to entering a panic state, Alice
could vote more than one time using to ensure that
she entered the password correctly at least once. However this is a specific
example and we would like a solution for the entire class.
If we generalize P-Compliment, Alice
knows two things: one real password and a rule, `anything
except
is a panic
password.' The fundamental problem is that the rule
is too inclusive. We can generalize to four desirable properties,
1. an arbitrarily large number of panic passwords,
2. an arbitrarily large set of invalid passwords,
3. a small distance between panic and invalid passwords,
4. a rule that is cognitively easy to apply.
The reason for the first property is
twofold: we wish to prevent an iteration attack without locking the account,
and as a password scheme, it must have all the security properties of any
password system. As such, it must not be susceptible to exhaustive search
attacks. The second property is to promote the probability that a mistyped is an invalid
password instead of a panic password. This property is necessary but not
sufficient---we also need to assure that panic and invalid password spaces are
well mixed. For example, a good metric for passwords is the Damerau-Levenshtein
distance [2], which is the minimum number of a specific set of operations to
convert one string into another, where the operations are insertion, deletion,
or substitution of a single character and transposition of two characters. The
metric is designed for application to spell-checking. Finally, from a usability
perspective, users need to apply the rule confidently, especially since they
are operating under duress.
As a general approach, we start with a
password space and
we apply the separating function
to each
element in it such that any
is mapped to
or
(i.e.,
valid or invalid).
and
should both be
sufficiently large.
is selected
from
and a panic
password is
. All elements
of
are invalid.
One implementation of this general
scheme is used by some US governmental agencies [3]. A password is combined
with a PIN to create . Panic
passwords are any combination such that the password portion is correct and the
PIN is incorrect. Compositions of a wrong password are considered invalid. This
scheme is problematic if a mistake is made on the PIN portion. Also, under
coercion, Alice is forced to reveal the password part of
in order to
use a
which reduces
the security of
against an
exhaustive search significantly. Since the threat model these passwords are
used under is unknown by the authors, we are not suggesting this scheme is used
improperly---only that it has drawbacks for scenarios of this type.
5-Dictionary. Let be
any five strings and let
be any five
strings in a standard dictionary. Alice composes a password by selecting a set
of five dictionary words. Any other set of five dictionary words is a panic
password. Any other set of either dictionary or non-dictionary strings is
considered invalid.
This systems meets the five criteria listed above. The standard Unix dictionary is just over 25,000 words, which provides up to 73 bits of security for a 5 word combination. The invalid password space is arbitrarily large, as any word can be any other combination of alphanumerical characters. Although empirical data of common typos, along with numerical analysis would need to confirm the average Damerau-Levenshtein distance between typos and words, our intuition is that there is a considerable probability that many typos would be caught---certainly higher than in P-Compliment. A similar approach could be taken with a click-based graphical password scheme.
5-Click. Alice is presented with a sequence of five images, and a rule for
what regions of the image are in based on the
content of the picture. She knows one password
which is a
sequence of one click per image which she uses for normal authentication purposes.
To communicate panic, she clicks on another valid region for at least one of
her clicks.
For example, in the PassPoints scheme
[13], users authenticate by clicking on certain pixels (within a tolerance) in
an image. A separating function could be used to predefine certain regions of
an image for and give them
semantic meaning to help users remember them. For example,
could be faces on the cover image of Sgt. Pepper or cars in an
image of a parking lot. Using a sequence of different images will increase
entropy, and is similar to the Cued Click-Points (CCP) scheme [5], where users
authenticate by clicking one point in each of a sequence of five images.
However in CCP, the
image
displayed to the user is dependent on where the user clicked in the
image. This
last property has significant usability improvements, as it provides feedback
to the user in case of an error, but in this model it could also be used by an
adversary to screen---the user would know the sequence of images for
and but not
(unless she
explicitly committed a panic sequence to memory). Instead, using the same
sequence of images would be a suitable compromise between increased entropy and
usability.
For a final scenario, consider a
scenario with in a screening
model where persistent Oscar's goal is
. In
particular, our focus is on an inherent problem between
and screening.
Our approach to this scenario applies equivalently to
and
.
Working Example: Consider the scenario where Alice has network access to sensitive information, such as credit card account numbers, which are password protected. Under coercion, she wants the server to return false information that looks real or be routed through to a ``honeypot'' server [9].
Consider using 2P. Given that , Oscar can
apply the iteration principle to
instead of
. He can request that Alice authenticates to the server using a
different password such that it produces a different
. Once he observes two different
values, he knows one must be
. The solution
to this problem parallels the dictionary solution---the set of
must be arbitrarily large. For this reason, it is not always
possible to solve this problem. The password-side of scheme can be handled by
5-Dictionary but the response-side would have to contain randomization of data.
It may appear possible to compromise by
having a finite but secrete number of panic responses, such that Alice could
claim that all possible responses have been iterated through when at least the
real one still remains unseen by Oscar. There are two problems with this
approach. First, it is a slight breach of assumption 1 depending on how you
define what is part of the system and what is the shared secret. Second, if the
panic state draws randomly from a finite set of possibilities, these states
will eventually repeat while the legitimate data will not repeat.[2] For this reason, Oscar can screen out sets of data as illegitimate
once they appear more than once and use this property to force Alice to
authenticate with .
We leave a generalized solution to scenarios of this nature for future work. Presenting randomized responses that appear legitimate depends on the what ``appear legitimate'' and are thus situational. It appears that solutions would have to be tailored for the specific circumstances.
In the case of panic passwords, the ability to be issued a password or reset a password on the basis of supplemental authentication information provides a backdoor that an adversary can successfully exploit. Instead of forcing Alice to authenticate, Oscar would simply force her to reveal the information necessary to reset the password or be issued a new one. Without completely solving the problem, passwords could be issued but not reset, effectively limiting Oscar's window of opportunity---this may feasible in less critical scenarios. For more serious scenarios, there appears to be no alternative to having the user create and reset passwords in a controlled, coercion-free environment. This could be by visiting a physical bank to set an ATM PIN or online banking account, or a one-time in-person registration to enable online voting, which is likely an improvement over voting in person in every election on a specific day. In some countries, government services are provided under federated access and so the registration could be used in scenarios beyond voting.
When introducing the three participants in the threat model---Alice, Bob, and Oscar---we assumed Bob, the entity being authenticated to by Alice, was trustworthy. In certain scenarios, this is not an adequate assumption. For example, we considered the example of Internet-based voting where Bob's hidden reaction was to dispose of votes cast under a panic password and count only the votes cast under a normal password. Much voting technology research is aimed at eliminating the trust of any entity during the voting process, including the server. End-to-end (E2E) voter verifiability means that voters can verify that their cast ballots are included in the final tally unmodified. To take one example, the Scantegrity II voting system [12] adds E2E integrity to optical-scan election systems.[3] Like most E2E systems, Scantegrity II issues a receipt to a voter that allows her to ensure that her vote was cast-as-intended without revealing how she voted. It is not immediately apparent how receipt-based voting could be made compatible with panic password schemes. If Alice does not trust Bob, she wants proof (i.e., through a receipt and audit) that her real ballot was cast and her panic ballots were discarded. However such proof could also be used by Oscar to screen the ballots he forced her to cast. We leave this problem open to future work.
Despite a large volume of research on the use of passwords, examined from many diverse angles---security, usability, agreement schemes, storage, dual factor, biometrics, to name a few---academic study of panic passwords is virtually non-existent. While they appear in some commercial and military applications, our hope is that this paper provides a basis for further attention from the research community. We believe that better schemes are possible and more involved threat models could be devised. In particular, attention from a usability perspective would be extremely valuable, including user studies and case studies.
[1] Ross J. Anderson and Roger M. Needham and Adi Shamir. The Steganographic File System. Proceedings of the Second International Workshop on Information Hiding, pages 73--82, 1998.
[2] Bard, Gregory V. Spelling-Error Tolerant, Order-Independent Pass-Phrases via the Damerau-Levenshtein String-Edit Distance Metric. Fifth Australasian Information Security Workshop (AISW 2007), pages 117-124, 2007.
[3] Richard T. Carback. Private conversations. April 2008.
[4] David Chaum. Achieving Electronic Privacy. Scientific American, :96--101, 1992.
[5] Sonia Chiasson and Paul C. van Oorschot and Robert Biddle. Graphical Password Authentication Using Cued Click Points. ESORICS in Lecture Notes in Computer Science, pages 359-374, 2007. Springer.
[6] Leemon C. Baird, III and Mance E. Harmon and R. Reed Young and James E. Armstrong, Jr. Apparatus and method for authenticating access to a network resource. United States Patent, 6732278, 2004.
[7] Ronald J. Massa and Theodore R. Ellis and Robert G. LePage. Intelligent surveillance alarm system and method. United States Patent, 4589081, 1986.
[8] A Munje and T Plestid. Password methods and systems for use on a mobile device. United States Patent (Pending), 11181522, 2007.
[9] Niels Provos. A virtual honeypot framework. SSYM'04: Proceedings of the 13th conference on USENIX Security Symposium, pages 1--14, 2004.
[10] Ronald K. Russikoff. Computerized password verification system and method for ATM transactions. United States Patent, 6871288, 2005.
[11] FIPR Response to the Home Office: ``Consultation on the Draft Code of Practice for the Investigation of Protected Electronic Information --– Part III of the Regulation of Investigatory Powers Act 2000''. http://www.fipr.org/060901encryption.pdf, September 2006.
[12] David Chaum and Richard Carback and Jeremy Clark and Aleksander Essex and Stefan Popoveniuc and Ronald L. Rivest and Peter Y. A. Ryan and Emily Shen and Alan T. Sherman. Scantegrity II: End-to-End Verifiability for Optical Scan Election Systems using Invisible Ink Confirmation Codes. EVT'08: Proceedings of the USENIX/Accurate Electronic Voting Technology Workshop, 2008.
[13] Susan Wiedenbeck and Jim Waters and Jean-Camille Birget and Alex Brodskiy and Nasir Memon. PassPoints: design and longitudinal evaluation of a graphical password system. Int. J. Hum.-Comput. Stud., 63(1-2):102--127, 2005.
[1]To demonstrate why Oscar's goal
is and not
, consider a situation where
Oscar forces Alice to authenticate twice, once yielding
and once
. This outcome is unacceptable to
Oscar if he has the former goal but is acceptable for the latter goal. If
is a silent alarm, this outcome
is unacceptable and thus Oscar must not have the latter goal in this scenario.
[2]We assume a consistent username is used by Alice.
[3]In optical-scan systems, voters mark a paper ballot by penciling in an oval and this ballot is then scanned for electronic tallying.