0% found this document useful (0 votes)
80 views8 pages

Anon V Panopticon PDF

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
80 views8 pages

Anon V Panopticon PDF

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Seeking Anonymity in an Internet Panopticon

UNPUBLISHED DRAFT

Joan Feigenbaum and Bryan Ford


Yale University

1. INTRODUCTION through software exploits or user error, an attacker can often cir-
In today’s “Big Data” Internet, users often need to assume that, cumvent anonymity tools entirely [24].
arXiv:1312.5307v3 [cs.CR] 3 Jan 2015

by default, their every statement or action online is monitored and Current approaches to anonymity also appear unable to offer ac-
tracked. Users’ statements and actions are routinely linked with curate, principled measurement of the level or quality of anonym-
detailed profiles built by entities ranging from commercial vendors ity a user might obtain. Considerable theoretical work analyzes
and advertisers to state surveillance agencies to online stalkers and onion routing [10], but relies on idealized formal models making
criminal organizations. Indeed, recent events have raised the stakes assumptions that are unenforceable and may be untrue in real sys-
in Internet monitoring enormously. Documents leaked by Edward tems – such as that users choose relays and communication partners
Snowden have revealed that the US government is conducting war- at random – or depending on parameters that are unknown in prac-
rantless surveillance on a massive scale and, in particular, that the tice, such as probability distributions representing user behavior.
long-term goal of the National Security Agency is to be “able to We believe the vulnerabilities and measurability limitations of
collect virtually everything available in the digital world” [18]. onion routing may stem from an attempt to achieve an impossi-
Internet users often have legitimate need to be anonymous – i.e., ble set of goals and to defend an ultimately indefensible position.
“not named or identified” by Webster’s definition of the term – to Current tools offer a general-purpose, unconstrained, and individ-
protect their online speech and activities from being linked to their ualistic form of anonymous Internet access. However, there are
real-world identities. Although the study of anonymous-communi- many ways for unconstrained, individualistic uses of the Internet to
cation technology is often motivated by high-stakes use cases such be fingerprinted and tied to individual users. We suspect that the
as battlefield communication, espionage, or political protest against only way to achieve measurable and provable levels of anonymity,
authoritarian regimes, anonymity actually plays many well accepted and to stake out a position defensible in the long term, is to develop
roles in established democratic societies. For example, paying cash, more collective anonymity protocols and tools. It may be necessary
voting, opinion polling, browsing printed material in a book store or to constrain the normally individualistic behaviors of participating
library, and displaying creativity and low-risk experimentalism in nodes, the expectations of users, and possibly the set of applications
forums such as slashdot or 4chan are everyday examples of anony- and usage models to which these protocols and tools apply.
mous activity. Author JK Rowling used a pen name on a recent Toward this end, we offer a high-level view of the Dissent project,
post-Harry Potter novel, presumably not out of any fear of censor- a “clean-slate” effort to build practical anonymity systems embody-
ship or reprisal, but merely “to publish without hype or expectation ing a collective model for anonymous communication. Dissent’s
and . . . to get feedback under a different name” [22]. collective approach to anonymity is not and may never be a “drop-
Obtaining and maintaining anonymity on the Internet is chal- in” functional replacement for Tor or the individualistic, point-to-
lenging, however. The state of the art in deployed tools, such as point onion routing model it implements. Instead, Dissent sets out
Tor [1], uses onion routing (OR) to relay encrypted connections on to explore radically different territory in the anonymous-commu-
a detour passing through randomly chosen relays scattered around nication design space, an approach that presents advantages, disad-
the Internet. OR is scalable, supports general-purpose point-to- vantages, and many as-yet-unanswered questions. An advantage is
point communication, and appears to be effective against many of that the collective approach makes it easier to design protocols that
the attacks currently known to be in use [12]. Unfortunately, OR is provably guarantee certain well defined anonymity metrics under
known to be vulnerable to several classes of attacks for which no arguably realistic environmental assumptions. A disadvantage is
solution is known or believed to be forthcoming soon. For exam- that the collective approach is most readily applicable to multicast-
ple, via traffic confirmation, an attacker who compromises a major oriented communication, and currently much less efficient or scal-
ISP or Internet exchange might in principle de-anonymize many able than OR for point-to-point communication.
Tor users in a matter of days [14]. Through intersection attacks, Dissent follows in the tradition of Herbivore [20], the first at-
an adversary can rapidly narrow the anonymity of a target via ac- tempt to build provable anonymity guarantees into a practical sys-
tions linkable across time, in much the same way Paula Broadwell tem, and to employ dining cryptographers or DC-nets [5]. Dissent
and the “High Country Bandits” were de-anonymized [19]. Finally, utilizes both DC-nets and verifiable shuffles [17], showing for the
first time how to scale the formal guarantees embodied in these
This material is based upon work supported by the Defense Ad- techniques to offer measurable anonymity sets on the order of thou-
vanced Research Projects Agency (DARPA) and SPAWAR Sys- sands of participants [23]. Dissent’s methods of scaling individual
tems Center Pacific, Contract No. N66001-11-C-4018.
Permission to make digital or hard copies of all or part of this work for anonymity sets are complementary and synergistic with techniques
personal or classroom use is granted without fee provided that copies are Herbivore pioneered for managing and subdividing large peer-to-
not made or distributed for profit or commercial advantage and that copies peer anonymity networks; combining these approaches could en-
bear this notice and the full citation on the first page. To copy otherwise, to able further scalability improvements in the future.
republish, to post on servers or to redistribute to lists, requires prior specific Dissent incorporates the first systematic countermeasures to ma-
permission and/or a fee.
Copyright 2008 ACM 0001-0782/08/0X00 ...$5.00.
jor classes of known attacks, such as global traffic analysis and
intersection attacks [16, 25]. Because anonymity protocols alone
cannot address risks such as software exploits or accidental self-
identification, the Dissent project also includes Nymix, a proto-
type operating system that hardens the user’s computing platform
against such attacks [24]. Dissent and Nymix OS can of course
offer only network-level anonymity, in which the act of commu-
nicating does not reveal which user sent which message. No ano-
nymity system can offer users personal anonymity if, for example,
they disclose their real-world identities in their message content.
Figure 1: Onion routing (OR).
While at this time Dissent is still a research prototype not yet
ready for widespread deployment, and may never be a direct re-
placement for OR tools such as Tor because of possibly fundamen- communications around a distributed network of relays run by vol-
tal tradeoffs, we hope that it will increase the diversity of practical unteers all around the world; it prevents somebody watching your
approaches and tools available for obtaining anonymity online. Internet connection from learning what sites you visit, and it pre-
Section 2 presents the basics of OR and Tor. In Section 3, we vents the sites you visit from learning your [network] location.”
describe four problems with OR that have gone unsolved for many The project provides free application software that can be used for
years and may unfortunately be unsolvable. Section 4 provides an web browsing, email, instant messaging, Internet relay chat, file
overview of the Dissent approach to anonymous communication, transfer, and other common Internet activities; users can also ob-
and Section 5 contains open problems and future directions. tain free downloads that integrate the underlying Tor protocol with
established browsers, email clients, etc. Importantly, Tor users can
2. ONION ROUTING AND TOR easily (but are not required to) transform their Tor installations into
Tor relays, thus contributing to the overall capacity of the Tor net-
Currently the most widely deployed, general-purpose system for
work. Currently, there are approximately 40M “mean daily users”
anonymous Internet communication is Tor [1]. Tor’s technical foun-
of Tor worldwide, slightly over 10% of whom are in the United
dation is onion routing [13], derived in turn from mixnets [7].
States, and approximately 4700 relays. These and other statistics
Onion routing (OR) uses successive layers of encryption to route
are regularly updated on the Tor Metrics Portal [2].
messages through an overlay network, such that each node knows
The IP addresses of Tor relays are listed in a public directory so
the previous and the next node in the route but nothing else. More
that Tor clients can find them when building circuits. (Tor refers
precisely, let (V, E) be a connected, undirected network and R ⊆
to routes as “circuits,” presumably because Tor is typically used
V be a set of nodes serving as relays. The set R is known to all
for web browsing and other TCP-based applications in which traf-
nodes in V , as is the public key Kr , usable in some globally agreed-
fic flows in both directions between the endpoints.) Clearly, this
upon public-key cryptosystem, for each node r ∈ R. There is a
makes it possible for a network operator to prevent its users from
routing protocol that any node in V can use to send a message to
accessing Tor. The operator can simply disconnect the first hop in
any other node, but the nodes need not know the topology (V, E).
a circuit, i.e., the connection between the client and the first Tor
If node s wishes to send message M to node d anonymously,
relay, because the former is inside the network and the latter is out-
s first chooses a sequence (r1 , r2 , . . . , rn ) of relays. It then con-
side; this forces the Tor traffic to flow through a network gateway,
structs an “onion” whose n layers contain both the message and the
at which the operator can block it. Several countries that operate
routing information needed to deliver it without revealing node s’s
national networks, including China and Iran, have blocked Tor in
identity to any node except the first relay r1 . The core of the onion
precisely this way. Similarly, website operators can block Tor users
is (d, M ), i.e., the destination node and the message itself. The nth
simply by refusing connections from the last relay in a Tor circuit;
or innermost layer of the onion is
Craigslist is an example of a US-based website that does so. As a
On = (rn , ENCKrn (d, M )), partial solution, the Tor project supports bridges, or relays whose
IP addresses are not listed in the public directory, of which there are
i.e., the nth relay node and the encryption of the core under the nth currently approximately 2000. Tor bridges are just one of several
relay’s public key. More generally, the ith layer Oi , 1 ≤ i ≤ k − 1, anti-blocking or censorship-circumvention technologies.
is formed by encrypting the (i + 1)st layer under the public key of There is inherent tension in OR between low latency, one as-
the ith relay and then prepending the ith relay’s identity ri : pect of which is short routes (or, equivalently, low values of k),
Oi = (ri , ENCKri (Oi+1 )). and strong anonymity. Because its goal is to be a low-latency
anonymous-communication mechanism, usable in interactive, real-
Once it has finished constructing the outermost layer time applications, Tor uses 3-layer onions, i.e., sets k = 3 as in
O1 = (r1 , ENCKr1 (O2 )), Figure 1. Despite this choice of small k, many potential users re-
ject Tor because of its performance impact [8].
node s sends ENCKr1 (O2 ) to r1 , using the routing protocol of
the underlay network (V, E). When relay ri , 1 ≤ i ≤ n, re-
ceives ENCKri (Oi+1 ), it decrypts it using the private key kri 3. ATTACKS ON ONION ROUTING
corresponding to Kri , thus obtaining both the identity of the next We now summarize four categories of known attacks to which
node in the route and the message that it needs to send to this next OR is vulnerable and for which no general defenses are known.
node (which it sends using the underlying routing protocol). When
i = n, the message is just the core (d, M ), because, strictly speak- Global traffic analysis.
ing, there is no On+1 . We assume that d can infer from routing- OR was designed to be secure against a local adversary, i.e., one
protocol “header fields” of M that it is the intended recipient and that might eavesdrop on some network links and/or compromise
need not decrypt and forward. See Figure 1. some relay nodes but only a small percentage of each. It was not
Tor is a popular free-software suite based on OR. As explained designed for security against traffic analysis by a global adversary
on the Torproject website [1], “Tor protects you by bouncing your that can monitor large portions of the network constantly.
Figure 2: Traffic confirmation or fingerprinting to de-
anonymize onion-routing circuits
Figure 4: Example of an intersection attack

interspersed with a visit to another relay, as shown in Figure 3. Re-


gardless of how congestion is incurred, it slows all circuits passing
through this relay, including the victim circuit, if and only if that
circuit passes through the targeted relay. The attacker can there-
fore test whether a particular victim circuit flows through a par-
ticular router, simply by checking whether the victim circuit’s av-
erage throughput (which can be measured at any point along the
Figure 3: Example of a congestion-based active attack
circuit) slows down during the period of attacker-generated con-
gestion. The attacker repeatedly probes different relays this way
The most well known global-traffic-analysis attack—traffic con- until he identifies the victim’s entry and middle relays. Finally,
firmation—was understood by Tor’s designers but considered an the attacker might fully de-anonymize the user by focusing traffic
unrealistically strong attack model and too costly to defend against analysis on, or hacking, the user’s entry relay.
[1]. In the standard scenario illustrated in Figure 2, we assume that
the attacker cannot break Tor’s encryption but can monitor both the Intersection attacks.
encrypted traffic flowing from the user to the first or entry relay and In most practical uses of anonymous communication, a user typ-
the traffic flowing from the final or exit relay to the user’s commu- ically needs to send not just a single “one-off” message anony-
nication partner. This situation, while unlikely a decade ago, might mously but a sequence of messages that are explicitly related and
be realistic today if both the user and her communication target are hence inherently linkable to each other. For example, Tor clients
located in a single country, and the attacker is an ISP controlled need to maintain persistent TCP connections and engage in back-
or compromised by a state-level surveillance agency. In this case, and-forth “conversations” with web sites in order to support inter-
the attacker in principle need only monitor the entry and exit traffic active communication, sending new HTTP requests that depend on
streams and correlate them via known fingerprinting methods. the web server’s responses to the client’s previous HTTP requests.
For decades, this global-passive-adversary attack model was re- It is manifestly obvious at least to the web server (and probably to
garded as unrealistically strong, and used to justify “conservative” any eavesdropper who can monitor the connection between the Tor
assumptions in formal models [10]. Unfortunately, this adversar- exit relay and the web site) which packets comprise the same Web
ial model is now not only realistic but in fact too weak. With the communication session, even if it is not (yet) clear who initiated
commercialization and widespread deployment of routers that can that session. Further, if the user leaves an anonymous browser win-
perform deep packet inspection and modification, including “Man- dow open for an extended period or regularly logs into the same
in-the-Middle attacks” against encrypted SSL streams at line rate anonymous Web mail account, an eavesdropper may be able to link
[11], it has become clear that any realistic adversary must be as- many of the user’s browsing sessions together over a long period
sumed to be active, i.e., able to modify traffic streams at will. of time. Even if each message gives the attacker only a small and
statistically uncertain amount of information just slightly narrow-
Active attacks. ing the identity of the anonymous user, combining this information
The ability for an attacker to interfere actively in an anonymity across many observation points at different times rapidly strength-
network creates a wide array of new attacks as well as ways to ens the attacker’s knowledge.
strengthen existing traffic-analysis attacks. Figure 3 illustrates one In one example of this attack illustrated in Figure 4, an authori-
example of a congestion attack [9]. In this scenario, we assume tarian government compels its ISPs or cellular carriers to turn over
that the attacker can directly monitor only one hop of a Tor circuit, logs of which customers were online and actively using the network
e.g., the traffic from the exit relay to the target web server. The during which periods of time. An anonymous dissident posts blog
attacker in this case might be “in the network” or might simply entries to a pseudonymous blog at different points in time. Assume
own or have compromised the web server. The attacker wishes that the attacker controls none of the user’s onion relays. Nor does
to determine the set of relays through which a long-lived circuit he control the blog server; he merely observes the times at which
owned by a particular user passes. the blog entries appeared and the fact that the posts are manifestly
The attacker chooses one relay at a time from Tor’s public da- linkable to each other, and he can correlate this information with
tabase and remotely attempts to increase that relay’s load by con- the ISP logs. Perhaps the subject of the blog is official corruption
gesting it. For example, the attacker might simulate many ordinary in a particular city, enabling the authoritarian state to guess that the
Tor users to launch a denial-of-service attack on the relay. The at- dissident lives in that city and narrow attention to a small set of lo-
tacker can amplify his power by creating artificially long “flower- cal ISPs. The attacker merely retrieves the sets of users who were
petal” circuits that visit the target relay multiple times, each visit online at each time a blog post appeared and intersects those sets.
munication proceeds in synchronous rounds. In each round, each
of the n clients encrypts a single message under m concentric lay-
ers of public-key encryption, using each of the m shufflers’ public
keys, in a standardized order. All n clients send their ciphertexts
to the first shuffler, which holds the private key to the outermost
layer of encryption in all the clients’ ciphertexts. The first shuf-
fler waits until it receives all n clients’ ciphertexts, then unwraps
this outermost encryption layer, randomly permutes the entire set
of ciphertexts, and forwards the permuted batch of n ciphertexts
to the next shuffler. Each shuffler in turn unwraps another layer of
encryption, permutes the batch of ciphertexts, then forwards them
to the next shuffler. The final shuffler then broadcasts all the fully
Figure 5: Example of a software-exploit attack decrypted cleartexts to all potentially interested recipients.
In an “honest-but-curious” security model in which we assume
Although there may be many thousands of users online at each of each shuffler correctly follows the protocol (without, for example,
these posting times individually, all users other than the dissident in inserting, removing, or modifying any ciphertexts), the output from
question are likely to have gone offline during at least one of these the last shuffler offers provable anonymity among all non-colluding
times (because of normal churn – the partly random comings and clients, provided at least one of the shufflers keeps its random per-
goings of most users), allowing the attacker to eliminate them from mutation secret. Unfortunately, if any of the shufflers is actively
the victim’s anonymity set. The attacker simply needs to “wait and dishonest, this anonymity is easily broken. For example, if the first
watch” until the dissident has posted enough blog entries, and the shuffler duplicates the ciphertext of some attacker-chosen client,
intersection of the online-user sets will shrink to a singleton. then the attacker may be able to distinguish the victim’s cleartext
The strength of this attack in practice is amply demonstrated by in the shuffle’s final output simply by looking for the cleartext that
the fact that similar reasoning is used regularly in law enforce- appears twice in the otherwise-anonymized output batch.
ment [19]. The FBI caught a Harvard student who used Tor to A substantial body of work addresses these vulnerabilities to
post a bomb threat by effectively intersecting the sets of Tor users such active attacks. In a sender-verifiable shuffle [4, 6], each client
and Harvard network users at the relevant time. Paula Broadwell inspects the shuffle’s output to ensure that its own message was not
was de-anonymized via the equivalent of an intersection attack, as dropped, modified, or duplicated before allowing the shuffled mes-
were the “High Country Bandits”. Intersection attacks also form sages to be fully decrypted and used. More sophisticated and com-
the foundation of the NSA’s CO-TRAVELER program, which links plex provable shuffles, such as Neff’s [17], enable each shuffler to
known surveillance targets with unknown potential targets as their prove to all observers the correctness of its entire shuffle, i.e., that
respective cellphones move together from one cell tower to another. the shuffler’s output is a correct permutation of its input, without
revealing any information about which permutation it chose.
Software exploits and self-identification. Both types of verifiable shuffles offer cryptographic guarantees
No anonymous communication system can succeed if other soft- that the process of shuffling reveals no information about which of
ware the user is running gives away his network location. In a re- the n clients submitted a given message appearing in the shuffled
cent attack against the Tor network, illustrated in Figure 5, a num- output. Shuffling has the practical disadvantage that the level of
ber of hidden services (web sites whose locations are protected by security achievable against potentially compromised shufflers de-
Tor and which can be accessed only via Tor) were compromised pends on the number of shufflers in the path, and multiple shuf-
so as to send malicious JavaScript code to all Tor clients who con- flers must inherently be placed in sequence to improve security; in
nected to them. This malicious JavaScript exploited a vulnerabil- essence, latency is inversely proportional to security. The typical
ity in a particular version of Firefox distributed as part of the Tor cascade arrangement above, where all clients send their messages
Browser Bundle. This exploit effectively “broke out” of the usual through the same sequence of shufflers at the same time, is most
JavaScript sandbox and ran native code as part of the browser’s pro- amenable to formal anonymity proofs, but exacerbates the perfor-
cess. This native code simply invoked the host operating system to mance problem by creating the “worst possible congestion” at each
learn the client’s true (de-anonymized) IP address, MAC address, shuffler in succession instead of randomly distributing load across
etc., and sent them to an attacker-controlled server. many shufflers as an ad hoc, individualistic OR network would.
For these reasons, verifiable shuffles may be practical only when
4. COLLECTIVE ANONYMITY IN DISSENT high latencies are tolerable, and shufflers are well provisioned. One
relevant application is electronic voting, for which some shuffle
As a step toward addressing these challenges, we now intro-
schemes were specifically intended, and which might readily toler-
duce Dissent, a project that expands the design space and explores
ate minutes or hours of latency. A second application that arguably
starkly contrasting foundations for anonymous communication.
fits this model is anonymous remailers [7], which were popular be-
4.1 Alternative foundations for anonymity fore onion routing. Practical remailer systems have never to our
knowledge employed state-of-the-art verifiable shuffles featuring
Quantification and formal analysis of OR security under realis-
anonymity proofs, however, and were vulnerable to active attacks
tic conditions has proven an elusive goal [10]. Dissent therefore
analogous to the message duplication attack mentioned above.
builds on alternative anonymity primitives with more readily prov-
able properties: verifiable shuffles and dining cryptographers. Dining cryptographers.
The only well studied foundation for anonymity not based on se-
Verifiable shuffles. quential relaying is Dining Cryptographers or DC-nets, invented by
In a typical cryptographic shuffle, participating nodes play two
David Chaum in the late 1980s [5] but never used in practical sys-
disjoint roles: there is a set of n clients with messages to send and
tems until two decades later by Herbivore [20]. Instead of relaying,
a set of m shufflers that randomly permute those messages. Com-
While theoretically appealing, however, DC-nets have not been per-
ceived as practical, for at least three reasons illustrated in Figure 7.
First, in groups of size N , optimal security normally requires all
pairs of cryptographers to share coins, yielding complexity Ω(N 2 ),
both computational and communication. Second, large networks
of “peer-to-peer” clients invariably exhibit high churn, with clients
going offline at inopportune times; if a DC-nets group member dis-
appears during a round, the results of the round become unusable
and must be restarted from scratch. Third, large groups are more
likely to be infiltrated by misbehaving members who might wish to
block communication, and any member of a basic DC-nets group
Figure 6: The Dining Cryptographers approach to anonymous can trivially—and anonymously—jam all communication simply
communication. Alice reveals a 1-bit secret to the group, but by transmitting a constant stream of random bits.
neither Bob nor Charlie learn which of the other two members
sent this message. 4.2 Practical dining cryptographers
Utilizing the DC-nets foundation in practical systems requires
solving two main challenges: jamming and scalability. Herbivore [20]
pioneered the exploration of practical solutions to both of these
problems, and the Dissent project continues this work.

The jamming problem.


Both Chaum’s original paper [5] and many follow-up works stud-
ied theoretical solutions to the jamming problem, but were complex
and to our knowledge never put into practice. Herbivore sidestepped
the jamming problem by securely dividing a large peer-to-peer net-
work into many smaller DC-nets groups, enabling a peer who finds
himself in an unreliable or jammed group to switch groups until he
finds a functioning one. This design has the advantage of scaling to
support arbitrary-sized networks, with the downside that each peer
obtains provable anonymity only within his own group – typically
tens of nodes at most – and not guaranteeing anonymity within the
larger network. A second downside of switching groups to avoid
jamming is that an attacker who runs many Sybil nodes and selec-
tively jams only groups he cannot compromise completely, while
offering good service in groups in which he has isolated a single
Figure 7: Why DC-nets are hard to scale in practice: (1) worst-
“victim” node, can make it more likely that a victim “settles” in a
case N × N coin-sharing matrix; (2) network churn requires
compromised group than an uncompromised one [3].
rounds to start over; (3) malicious members can anonymously
Dissent, the only system since Herbivore to put DC-nets into
jam the group.
practice, explores different solutions to these challenges. First,
Dissent addresses the jamming problem by implementing account-
DC-nets build on information-coding methods. ability mechanisms, allowing the group to revoke the anonymity
Consider Chaum’s standard scenario, illustrated in Figure 6. Three of any peer found to be attempting to jam communication ma-
cryptographers are dining at a restaurant when the waiter informs liciously while preserving strong anonymity protection for peers
them that their meal has been paid for. Growing suspicious, they who “play by the rules.” Dissent’s first version introduced a con-
wish to learn whether one of their group paid the bill anonymously, ceptually simple and clean accountability mechanism that lever-
or NSA agents at the next table paid it. So each adjacent pair of aged the verifiable-shuffle primitive discussed above, at the cost of
cryptographers flips a coin that only the two can see. Each cryp- requiring a high-latency shuffle between each round of (otherwise
tographer XORs the coins to his left and right and writes the re- more efficient) DC-nets communication. The next version [23] in-
sult on a napkin everyone can see—except any cryptographer who troduced a more efficient but complex retroactive-blame mecha-
paid the bill (Alice in this case), who flips the result of the XOR. nism, allowing lower-latency DC-nets rounds to be performed “back-
The cryptographers then XOR together the values written on all the to-back” in the absence of jamming and requiring an expensive
napkins. Because each coin toss affects the values of exactly two shuffle only once per detected jamming attempt.
napkins, the effects of the coins cancel out of the final result, leav- An adversary who manages to infiltrate a group with many ma-
ing a 1 if any cryptographer paid the bill (and lied about the XOR) licious nodes, however, could still “sacrifice” them one-by-one to
or a 0 if no cryptographer paid. A 1 outcome provably reveals no create extended denial-of-service attacks. Addressing this risk, Dis-
information about which cryptographer paid the bill, however: Bob sent’s most recent incarnation [6] replaces the “coins” of classic
and Charlie cannot tell which of the other two cryptographers paid DC-nets with pseudorandom elliptic-curve group elements, replaces
it (unless of course they collude against Alice). the XOR combining operator with group multiplication, and re-
DC-nets generalize readily to support larger groups and trans- quires clients to prove their DC-nets ciphertexts correct on sub-
mission of longer messages. Typically each pair of cryptographers mission, using zero-knowledge proofs. To avoid the costs of us-
uses Diffie-Hellman key exchange to agree on a shared seed for a ing elliptic-curve cryptography all the time, Dissent implements a
standard pseudorandom-bit generator, which efficiently produces hybrid mode that uses XOR-based DC-nets unless jamming is de-
the many “coin flips” needed to anonymize multi-bit messages. tected, at which point the system switches to elliptic-curve DC-nets
Figure 8: Improving scalability and churn resistance through
an asymmetric, client/server DC-nets architecture.

only briefly to enable the jamming victim to broadcast an accusa-


tion, yielding a more efficient retroactive-blame mechanism.
Figure 9: Fingerprinting or staining attacks
Scaling and network churn.
Even with multiple realistic solutions to the jamming problem
now available, DC-nets cannot offer useful anonymity if they can sharing design addresses network churn by making the composition
guarantee anonymity-set sizes of at most tens of members. Herbi- of client ciphertexts independent of the set of other clients online in
vore addressed the N × N communication complexity problem via a given round. The servers set a deadline, and all clients currently
a star topology, in which a designated member of each group col- online must submit their ciphertexts by that deadline or risk being
lects other members’ ciphertexts, XORs them together, and broad- “left out” of the round. Unlike prior DC-nets designs, if some Dis-
casts the results to all members. Without a general solution to the sent clients miss the deadline, the other clients’ ciphertexts remain
network churn and jamming problems, however, both Herbivore usable. The servers merely adjust the set of client/server-shared
and the first version of Dissent were limited in practice to small secrets they use to compute their server-side DC-net ciphertexts.
anonymity sets comprising at most tens of nodes. Because each client’s ciphertext depends on secrets it shares with
To address churn and scale DC-nets further, Dissent now adopts all servers, no client’s ciphertext can be used or decrypted unless all
a client/multi-server model with trust split across several servers, servers agree on the same set of online clients in the round and pro-
preferably administered independently. No single server is trusted; duce correct server-side ciphertexts based on that agreement. Mali-
in fact, Dissent preserves maximum security provided only that not cious servers can at most corrupt a round and cannot de-anonymize
all of a group’s servers maliciously collude against their clients. clients except by colluding with all other servers.
The clients need not know or guess which server is trustworthy but
must merely trust that at least one trustworthy server exists. 4.3 How Dissent handles attacks
When a Dissent group is formed, the group creator defines both We now summarize how Dissent handles the attacks in Section 3.
the set of servers to support the group and the client-admission pol-
icy; in the simplest case, the policy is simply a list of public keys Global traffic analysis.
representing group members. Dissent servers thus play a role anal- Dissent builds on anonymity primitives that have formal secu-
ogous to relays in Tor, serving to support the anonymity needs of rity proofs in a model where the attacker is assumed to monitor all
many different clients and groups. Like Tor relays, the Dissent network traffic sent among all participating nodes but cannot break
servers supporting a new group might be chosen automatically from the encryption. We have extended these formal security proofs to
a public directory of available servers to balance load. Choosing the cover the first version of the full Dissent protocol [21], and formal
servers for each group from a larger “cloud” of available servers in analysis of subsequent versions is in progress. Although verifiable
this way in principle enables Dissent’s design to support an arbi- shuffles differ from DC-nets in their details, both approaches share
trary number of groups, but the degree to which an individual group one key property that enables formal anonymity proofs: All par-
scales may be more limited. If a particular logical group becomes ticipants act collectively under a common “control plane” rather
extremely popular, Herbivore’s technique of splitting a large group than individually as in an ad hoc OR system. For example, they
into multiple smaller groups may be applicable. Our current Dis- send identical amounts of network traffic in each round, although
sent prototype does not yet implement either a directory service or amounts and allocations may vary from round to round.
Herbivore-style subdivision of large networks, however.
While individual groups do not scale indefinitely, Dissent ex- Active attacks.
ploits its client/multi-server architecture to make groups scale two One countermeasure to traffic analysis in OR is to “pad” connec-
orders of magnitude beyond prior DC-nets designs [23]. As illus- tions to a common bit rate. While padding may limit passive traffic
trated in Figure 8, clients no longer share secret “coins” directly analysis, it often fails against active attacks, for reasons illustrated
with other clients but only with each of the group’s servers. Since in Figure 9. Suppose a set of OR users pad the traffic they send to
the number of servers in each group is typically small (e.g., 3–5, a common rate, but a compromised upstream ISP wishes to “mark”
comparable to the number of Tor relays supporting a circuit), the or “stain” each client’s traffic by delaying packets with a distinctive
number of pseudorandom strings each client must compute is sub- timing pattern. An OR network, which handles each client’s circuit
stantially reduced. This change does not reduce anonymity, how- individually, preserves this recognizable timing pattern (with some
ever, subject to Dissent’s assumption that at least one server is hon- noise) as it passes through the relays, at which point the attacker
est. Chaum’s DC-nets security proof [5] ensures ideal anonymity might recognize the timing pattern at the egress more readily than
provided all honest nodes are connected via the coin-sharing graph; would be feasible with a traffic-confirmation attack alone. Active
Dissent satisfies this requirement, because the one honest server as- attacks also need not mark circuits solely via timing. A sustained
sumed to exist shares coins directly with all honest clients. attack deployed against Tor last year exploited another subtle pro-
More importantly in practice, Dissent’s client/multi-server coin- tocol side-channel to mark and correlate circuits, going undetected
for five months before being discovered and thwarted last July.
The collective-anonymity primitives underlying Herbivore and
Dissent, in contrast, structurally keep the clients comprising an
anonymity set in “lock-step,” under the direction of a common,
collective control plane. As in the popular children’s game “Si-
mon Says,” participants transmit when and how much the collec-
tive control plane tells them to transmit. A client’s network-visible
communication behavior does not leave a trackable fingerprint or
stain, even under active attacks such as those above, because its
network-visible behavior depends only on this anonymized, collec-
tive control state; that is, a client’s visible behavior never depends
directly on individual client state. Further, the Dissent servers im-
plementing this collective control plane do not know which user
owns which pseudonym or DC-nets transmission slot and thus can-
not leak that information via their decisions, even accidentally.
Contrary to the intuition that defense against global traffic anal- Figure 10: Nymix: using per-pseudonym virtual machines or
ysis and active attacks require padding traffic to a constant rate, NymBoxes to harden the client operating system against soft-
Dissent’s control plane can adapt flow rates to client demand by ware exploits, staining, and self-identification
scheduling future rounds based on (public) results from prior rounds.
For example, the control-plane scheduler dynamically allocates DC-
nets transmission bandwidth to pseudonyms who in prior rounds
anonymously indicated a desire to transmit and hence avoids wast-
Software exploits and self-identification.
ing network bandwidth or computation effort when no one has any-
No anonymity protocol, by itself, can prevent de-anonymization
thing useful to say. Aqua, a recent project to strengthen OR secu-
via software exploits or user self-identification. Nevertheless, the
rity, employs a similar collective-control philosophy to normalize
Dissent project is exploring system-level solutions to this problem
flow rates dynamically across an anonymity set [15]. In this way,
via Nymix, a prototype USB-bootable Linux distribution that em-
a collective control plane can in principle not only protect against
ploys virtual machines (VMs) to improve resistance to exploits [24].
both passive and active attacks but, ironically, can also improve ef-
As shown in Figure 10, Nymix runs anonymity-client software
ficiency over padding traffic to a constant bit rate.
(currently either Tor or Dissent) in the platform’s host operating
system but isolates the browser and any plug-ins and other exten-
Intersection attacks.
sions it may depend on in a separate Guest VM. No software in this
While the power and generality of intersection attacks has been
guest VM is given access to information about the physical host OS
extensively studied in the past decade, there has been scant work on
or its network configuration. For example, the guest VM sees only
actually building mechanisms to protect users of practical systems
a standard private (NATted) IP address such as 192.168.1.1 and the
against intersection attacks. The nearest precedents we are aware of
fake MAC address of a virtual device. Even native code injected
are suggestions that traffic padding may make intersection attacks
by the recent Tor Browser Bundle exploit would thus not be able to
more difficult [16]. To the best of our knowledge, such proposals
“leak” the client’s IP address without also breaking out of the VM
have never been implemented, in part because there is no obvious
(which of course may be possible, but raises the attack difficulty).
way to measure how much protection against intersection attacks a
Nymix binds guest-VM state instances to pseudonyms managed
given padding scheme will provide in a real environment.
by the anonymity layer, enabling users to launch multiple simulta-
Dissent is the first anonymity system designed with mechanisms
neous pseudonyms in different VMs or NymBoxes. Nymix securely
both to measure potential vulnerability to intersection attacks, us-
discards all pseudonym state embodied in a NymBox when desired
ing formally grounded but plausibly realistic metrics, and to of-
to minimize the user’s long-term exposure to intersection attacks.
fer users active control over anonymity loss under intersection at-
This binding of pseudonyms to VMs makes it easy for the user to
tacks [25]. Dissent implements two different anonymity metrics:
maintain state related to the context of one logical pseudonym (such
possinymity, a possibilistic measurement of anonymity-set size mo-
as Web cookies, open logins, etc.), while offering stronger pro-
tivated by “plausible-deniability” arguments, and indinymity, an in-
tection against the user’s accidentally linking different pseudonym
distinguishability metric effective against stronger adversaries that
VMs, because they appear as entirely separate OS environments
may make probabilistic “guesses” via statistical disclosure [16].
and not just different browser windows or tabs.
Users may set policies for long-lived pseudonyms limiting the
To reduce the risk of self-identification, Nymix allows the user
rate at which measured possinymity or indinymity may be lost, or
to “move” data between non-anonymous contexts, such as personal
setting a threshold below which these metrics must not fall. Dis-
JPEG photos stored on the host OS, and pseudonym-VM contexts
sent’s collective control plane enforces these policies in essence by
only via a quarantine file system “drop box.” Any files the user
detecting when allowing a communication round to proceed might
moves across browsing contexts in this way undergoes a suite of
reduce a pseudonym’s possinymity or indinymity “too much” and,
tests for possibly compromising information, such as EXIF meta-
in response, suppressing or delaying communication temporarily.
data within JPEGs. The quarantine system warns the user of any
The control plane can compute these metrics and enforce these
detected compromise risks and gives him the opportunity to scrub
policies even though its logic does not “know” which user actu-
the file or decide not to transfer it at all. While all of these de-
ally owns each pseudonym. The downside is that employing these
fenses are inherently “soft,” because there is only so much we can
controls to resist intersection attacks can reduce the responsiveness,
do to prevent users from shooting themselves in the foot, Nymix
availability, and/or lifetime of a pseudonym. We believe this cost
combines these VM-based isolation and structuring principles in
reflects a fundamental tradeoff between anonymity and availability.
an effort to make it easier for users to make appropriate and well
informed uses of today’s and tomorrow’s anonymity tools.
5. CHALLENGES AND FUTURE WORK Verdict. In 22nd USENIX Security Symposium, August 2013.
Dissent takes a few steps in developing a collective approach to [7] George Danezis, Roger Dingledine, and Nick Mathewson.
anonymous communication, but many practical challenges remain. Mixminion: Design of a Type III anonymous remailer
protocol. In IEEE Security and Privacy (SP), pages 2–15,
First, while DC-nets now scale to thousands of users, they need May 2003.
to scale to hundreds of thousands or more. One approach is to [8] Roger Dingledine and Steven J. Murdoch. Performance
combine Dissent’s scaling techniques with those of Herbivore [20] improvements on Tor or, why Tor is slow and what we’re
by dividing large anonymity networks into manageable anonym- going to do about it. In DEFCON 17, July 2009.
ity sets (e.g., hundreds or thousands of nodes), balancing perfor- [9] Nathan S. Evans, Roger Dingledine, and Christian Grothoff.
mance against anonymity guarantees. A second approach is to use A practical congestion attack on Tor using long paths. In
small, localized Dissent clusters, which already offer performance 18th USENIX Security Symposium, August 2009.
[10] Joan Feigenbaum, Aaron Johnson, and Paul Syverson.
adequate for interactive Web browsing [23, 24], as a decentralized Probabilistic analysis of onion routing in a black-box model.
implementation for the crucial entry-relay role in a Tor circuit [1]. ACM Transactions on Information and System Security,
Much of a Tor user’s security depends on his entry relay’s being 15(3):14:1–14:28, November 2012.
uncompromised [14]; replacing this single point of failure with a [11] Ryan Gallagher. New Snowden documents show NSA
Dissent group could distribute the user’s trust among the members deemed Google networks a “target”. Slate, September 9,
of this group and further protect traffic between the user and the 2013.
Tor relays from traffic analysis by “last mile” ISP adversaries. [12] Barton Gellman, Craig Timberg, and Steven Rich. Secret
Second, while Dissent can measure vulnerability to intersection NSA documents show campaign against Tor encrypted
network. The Washington Post, October 4, 2013.
attack and control anonymity loss [25], it cannot also ensure avail-
[13] David M. Goldschlag, Michael G. Reed, and Paul F.
ability if users exhibit high churn and individualistic, “every user Syverson. Hiding Routing Information. In 1st International
for himself” behavior. Securing long-lived pseudonyms may be Workshop on Information Hiding, May 1996.
feasible only in applications that incentivize users to keep commu- [14] Aaron Johnson, Chris Wacek, Rob Jansen, Micah Sherr, and
nication devices online consistently, even if at low rates of activ- Paul Syverson. Users get routed: Traffic correlation on Tor
ity, to reduce anonymity decay caused by churn. Further, robust by realistic adversaries. In 20th ACM Conference on
intersection-attack resistance may be practical only in applications Computer and Communications Security (CCS), November
2013.
designed to encourage users to act collectively, rather than individ-
[15] Stevens Le Blond, David Choffnes, Wenxuan Zhou, Peter
ually, and optimized for these collective uses. Druschel, Hitesh Ballani, and Paul Francis. Towards efficient
Applications in which users cooperatively produce collective in- traffic-analysis resistant anonymity networks. In ACM
formation “feeds” consumed by many others users may be well SIGCOMM, August 2013.
suited to Dissent’s collective anonymity model: e.g., the interac- [16] Nick Mathewson and Roger Dingledine. Practical traffic
tion models of IRC, forums like Twitter or Slashdot, or applications analysis: extending and resisting statistical disclosure. In 4th
supporting voting, deliberating, or “town hall” meetings. Given the International Workshop on Privacy Enhancing Technologies
(PETS), May 2004.
close relationship between collective deliberation and the founda-
[17] C. Andrew Neff. A verifiable secret shuffle and its
tions of democracy and freedom of speech, such applications may application to e-voting. In 8th ACM Conference on Computer
also represent some of the most socially important use cases for and Communications Security (CCS), November 2001.
online anonymity. How best to support and incentivize cooperative [18] James Risen and Laura Poitras. NSA report outlined goals
behavior, however, remains an important open problem. for more power. The New York Times, November 22, 2013.
Finally, it is clear that large anonymity sets require widespread [19] Aaron Segal, Bryan Ford, and Joan Feigenbaum. Catching
public demand for anonymity. Tor’s 40M “mean daily users” are bandits and only bandits: Privacy-preserving intersection
dwarfed in number by the users of Google, Facebook, Yahoo!, and warrants for lawful surveillance. In 4th USENIX Workshop
other services that do not provide anonymity – and cannot provide on Free and Open Communications on the Internet
(FOCI’14), August 2014.
it, because their business models depend crucially on exploitation [20] Emin Gün Sirer, Sharad Goel, Mark Robson, and Dǒgan
of personal information. Public demand for anonymity online may Engin. Eluding carnivores: File sharing with strong
rise as a result of the ongoing surveillance scandal, thereby provid- anonymity. In 11th ACM SIGOPS European Workshop,
ing an opportunity to deploy new anonymity tools. September 2004.
[21] Ewa Syta, Aaron Johnson, Henry Corrigan-Gibbs, Shu-Chun
Weng, David Isaac Wolinsky, and Bryan Ford. Security
6. REFERENCES analysis of accountable anonymity in Dissent. ACM
[1] Tor: Anonymity online. Transactions on Information and System Security (TISSEC),
https://www.torproject.org. 17(1), August 2014.
[2] Tor metrics portal. [22] Robert Watts. JK Rowling unmasked as author of acclaimed
http://metrics.torproject.org/. detective novel. The Telegraph, July 13, 2013.
[3] Nikita Borisov, George Danezis, Prateek Mittal, and Parisa [23] David Isaac Wolinsky, Henry Corrigan-Gibbs, Aaron
Tabriz. Denial of service or denial of security? How attacks Johnson, and Bryan Ford. Dissent in numbers: Making
on reliability can compromise anonymity. In Proceedings of strong anonymity scale. In 10th USENIX Symposium on
the 14th ACM Conference on Computer and Operating Systems Design and Implementation (OSDI),
Communications Security. October 2012.
[4] Justin Brickell and Vitaly Shmatikov. Efficient [24] David Isaac Wolinsky, Daniel Jackowitz, and Bryan Ford.
anonymity-preserving data collection. In 12th ACM SIGKDD Managing NymBoxes for identity and tracking protection. In
International Conference on Knowledge Discovery and Data USENIX Conference on Timely Results in Operating
Mining (KDD), August 2006. Systems, October 2014.
[5] David Chaum. The Dining Cryptographers problem: [25] David Isaac Wolinsky, Ewa Syta, and Bryan Ford. Hang with
Unconditional sender and recipient untraceability. Journal of your buddies to resist intersection attacks. In 20th ACM
Cryptology, pages 65–75, January 1988. Conference on Computer and Communications Security
[6] Henry Corrigan-Gibbs, David Isaac Wolinsky, and Bryan (CCS), November 2013.
Ford. Proactively accountable anonymous messaging in

You might also like