![]() ![]() | ||||||||||||||
|
![]() |
|
![]() |
This section describes Dandelion's cryptographic fair-exchange-based protocol.
By
we denote the description of an entity or object,
e.g.
denotes a client
's Dandelion ID.
is
's master
secret key,
is a cryptographic hash function
such as SHA-1,
is
a Message Authentication Code such as HMAC [20], and
refers to a time
period. By
we denote
at client or server
.
Due to host mobility and NATs, we do not use Internet address (IP or
IP/source-port) to associate credit and other persistent
protocol information with clients. Instead, each user applies for a
Dandelion account and is associated with a persistent ID. The server
associates each client with its authentication information (client ID
and password), the content (e.g. a file)
it currently
downloads or seeds, its credit balance, and the content it can access.
The clients and the server maintain loosely synchronized clocks.
Every client that wishes to join the network
must establish a
transport layer secure session with the server
, e.g. using TLS
[1]. A client sends
its ID and password over the secure channel. The server
generates a secret key and symmetric
encryption initialization vector pair, denoted
, which is
shared with
.
is efficiently
computed as
.
is also sent
over the secure channel.
This key is used both for symmetric encryption and for computing a
MAC. For MAC computation, we use only the secret key portion of
.
The rest of the messages that are exchanged between the server and the
clients are sent over an insecure communication channel (e.g. plain
TCP), which must originate from the same IP as the secure session.
Similarly, all messages between clients are sent over an insecure
communication
channel.
Each client exchanges only short messages with
the server. To prevent forgery of the message source and replay
attacks, and to ensure the integrity of the message, each message
includes a sequence number and a digital signature. The signature is
computed as the MAC of the message, keyed with the secret key
that
shares with the server. Each time a
client or the server receive a message from each other, they check
whether the sequence number succeeds the sequence number of the
previously received message and whether the MAC-generated signature
verifies. If either of the two conditions is not satisfied, the message
is discarded. The sequence number is reset when time period
changes.
The server initiates re-establishment of shared keys
with the clients upon
change in order to: a) prevent attackers from inferring
by examining the encrypted content and the MACs used by the protocol;
and b) allow the reuse of message sequence numbers once the numbers
reach a high threshold, while preventing attackers from replaying
previously signed and sent messages.
tolerates some lag in the
assumed by a client.
To provide robust incentives for cooperation under the model described
in Section 3.2,
Dandelion employs a cryptographic fair-exchange mechanism. Our
fair-exchange protocol involves only efficient
symmetric cryptographic operations. The server acts as the trusted
third party (TTP) mediating the exchanges of content for credit among
its clients. When a client
uploads to a client
, it sends encrypted content
to client
. To decrypt,
must request the decryption key
from the server. The requests for keys serve as the proof that
has uploaded some content to
. Thus, when the server receives a key request, it credits
for uploading content to
, and charges
for downloading
content.
When a client sends invalid content to
,
can determine that the content is invalid only after receiving
the decryption key and being charged. To address this problem, our
design includes a non-repudiable complaint mechanism. If
intentionally sends garbage to
,
cannot deny that it did. In
addition,
is prevented from falsely claiming
that
has sent it
garbage.
The following description omits the sequence number and the signature
in the messages between clients and the server. Figure 2 depicts the message
flow in our protocol.
Step 1:
The protocol starts with the client
sending a request for the content item
to
.
:
Step 2: If
has access to
,
chooses a short list of clients
,
which are currently in the swarm for
. The policy
with which the server selects the list
depends
on the specifics of the content distribution system. Each list entry,
besides the ID of the client, also contains the client's inbound
Internet address. For every client in
,
sends a ticket
to
.
is a timestamp,
and
is a client in
. The tickets
are only valid
for a certain amount of time
(considering clock skew between
and
) and allow
to request
chunks of the content
from client
. When
expires and
still wishes to download from
, it requests a new
from
.
To ensure integrity in the case of static content or video on demand,
also sends to
the SHA-1 hash
for all chunks
of
. For the case of live streaming content, the content provider augments
the
chunks it generates with his public key signature on their hash and ID,
as
. Clients append this signature to all the chunks they upload.
:
Step 3: The
client forwards this request to each
.
:
Step 4: If current-time and
is not in
's cache,
verifies if
.
The purpose of this check is to mitigate DoS attacks against
; it allows
to filter out requests from clients
that are not authorized to retrieve the content or from clients that
became blacklisted. As long as
remains connected to
, it periodically renews its
tickets by requesting them from
. If the verification fails,
drops this request. Also, if
is greater than
's current epoch
,
learns that it should renew its key
with
soon. Otherwise,
caches
and periodically
sends the chunk announcement message described below,
for as long as the timestamp
is fresh. This message contains a list of chunks that
owns,
.
also does so in separate chunk
announcement messages. The specifics
of which chunks are announced and how frequently depend on the type of
content
distribution. For example, in the case of static content distribution,
the initial chunk announcement would contain the IDs of all the chunks
owns, while the periodically sent announcement would contain the IDs of
newly acquired chunks.
:
Step 5:
and
determine which chunks to download
from each
other according to a chunk selection policy; BitTorrent's
locally-rarest-first is suitable for static content dissemination,
while for streaming content or video on demand
other policies are appropriate [42,23].
can request chunks from
, after it requests and
retrieves
from
.
sends a request for the missing
chunk
to
.
:
Step 6:
's chunk requests are served by
as long as the timestamp
is fresh, and
is cached or
verifies. If
is altruistic, it sends the chunk
to
in plaintext and the per-chunk
transaction ends here. Otherwise,
encrypts
using a symmetric encryption
algorithm
, as
.
is
a random secret key and random symmetric encryption initialization
vector pair. This pair is distinct for each chunk.
encrypts the random key with
, as
.
Next,
hashes the ciphertext
as
. Subsequently, it computes
its commitment to the encrypted chunk and the encrypted key as
and
sends the following to
.
:
Step 7: To
retrieve
,
needs to request it from the
server. As soon as
receives the encrypted chunk,
computes its own hash over the received ciphertext
and forwards the following to
.
:
Step 8: If
timestamp is fresh enough,
and
is not too much
off,
checks if
.
The time-stamp
freshness
requirement
forces
to expedite paying for decrypting
the encrypted chunks. This fact allows
to promply acquire credit for its service.
The ticket
verification may
fail either because
due to
transmission error in step
or because
or
are
misbehaving. Since
is unable to determine which is the case, it
punishes neither
or
and does not update their credit.
It does not send the decryption key to
but it still notifies
of the discrepancy. In this case,
is expected to disconnect from
and blacklist it in case
repeatedly
sends invalid chunk response messages. If
keeps sending invalid decryption key
requests,
penalizes him. If the verification
succeeds,
checks whether
has sufficient credit to purchase the chunk
.
It also checks again whether
has access to the content
. If
is
approved, it charges
and rewards
with
credit units.
Subsequently,
decrypts
,
and sends it to
.
:
uses
to decrypt the
chunk as
.
Next, we explain the complaint mechanism.
Step 9: If the
decryption fails or if
(step
),
complains to
by sending the following message.
:
ignores this message if current-time
,
where
.
should be greater than the time needed for
to receive a decryption key response, decrypt the chunk and send a
complaint to the server. With this condition, a misbehaving client
cannot avoid having complaints ruled against it, even if
ensures that the time elapsed between the moment
commits
to the encrypted chunk and the moment the encrypted chunk is received
by
is slightly less than
.
also ignores the complaint message
if a complaint for the same
and
is in a cache of recent complaints
that
maintains for each client
.
Complaints are evicted from this cache once current-time
.
If
,
punishes
. This is because
has already notified
in step
that
is invalid. If
verifies,
caches this complaint, recomputes
as before,
decrypts
once
again, retrieves
from its storage, and encrypts
himself using
,
.
If the hash of the ciphertext
is equal to the
value
that
sent to
,
decides that
has acted
correctly and
's complaint is unjustified. Subsequently,
drops the
complaint request and blacklists
. It also notifies
, which disconnects from
and blacklists it. Otherwise, if
,
decides that
was cheated by
, removes
from its set of active clients,
blacklists
, and revokes the corresponding credit charge on
. Similarly,
disconnects from
and blacklists it.
The server disconnects from a blacklisted client
, marks it as blacklisted in the credit file and denies access to
if it attempts to login. Future complaints concerning a blacklisted
client
and for which
verifies, are ruled against
without further processing.
Since a verdict on a complaint can adversely affect a client, each client needs to ensure that the commitments it generates are correct even in the rare case of a disk read error. Therefore, a client always verifies the read chunk against its hash before it encrypts the chunk and generates its commitment.
A content provider may be more concerned with scalability than it is with the free-riding problem presented in Section 2. In such case, it can deploy clients that use tit-for-tat incentives if their peers have content that interests them, i.e. the clients would upload plaintext content to peers that reciprocate with plaintext content. These clients would fall back to Dandelion's cryptographic fair-exchange mechanism when their peers do not have content that interests them. For example, selfish seeders would always upload encrypted content to their peers.
In case a client is unable to timely retrieve a missing chunk
from
its peers, it resorts to requesting the chunk from the server. If the
server is not busy, it replies with the plaintext chunk. If it is
moderately busy, it charges an appropriately large amount of credit
,
sends the chunk and indicates that it is preferable for the client
not to download chunks from the server. If the server is overloaded, it
replies with an error message. Clients always download the content from
the server in chunks, so that the system can seamlessly switch to the
peer-serving mode when
the server becomes busy.
Typically, a content distributor would deploy, in addition to the server, at least one client that possesses the complete content (initial seeder). In this way, the distributor ensures that the complete content is made available, even if the server is too busy to serve chunks.
This section briefly lists the security properties of
our design. For simplicity of presentation, we omit proofs on why
these properties hold. They can be found
in the Appendix of [51].
Lemma 4.1 If
the server charges a client
credit units for
a chunk
received from a selfish client
,
must have received the correct
, regardless of the actions taken by
.
Lemma 4.2 If
a selfish client
always encrypts chunk
anew when servicing a request and if
gets correct
from
, then
is awarded
credit units
from
, and
is charged
credit units
from
.
Lemma 4.3 A
selfish or a malicious client cannot assume another authorized client's
identity and issue messages under
, aiming at obtaining service at the expense of
, charging
for service it did not obtain or
causing
to be blacklisted. In addition, it
cannot issue a valid
for
an invalid chunk that it sends to a client
and cause
to produce a complaint message that
would result in a verdict against
.
Lemma 4.4 A
malicious client cannot replay previously sent valid requests to the
server or generate decryption key requests or complaints under
's ID, aiming at
being charged for service it did not obtain or being blacklisted
because of invalid or duplicate complaints.
Observation 4.5 A
client cannot download chunks
from a selfish peer if it does not have sufficient credit.
Our design choice to involve the server in every transaction, instead
of using the fair exchange technique proposed in [43],
enables the
server to check a client's credit balance before the client retrieves
the decryption
key of a chunk.
Observation 4.6 To
maintain an efficient content distribution pipeline, a client needs to
relay a received chunk to its peers as soon as it receives it. However,
the chunk may be invalid due to communication error or due to peer
misbehavior. The performance of the system would be severely degraded
if
peers wasted bandwidth to relay invalid content. To address this issue,
Dandelion
clients send a decryption key request to the server immediately upon
receiving the encrypted chunk. This design choice enables clients to
promptly retrieve the chunk in its non-encrypted form. Thus, they can
verify the chunk's integrity prior to uploading the chunk to their
peers.
Observation 4.7 A
malicious client cannot DoS attack the server by
sending invalid content to other clients or repeatedly sending invalid
complaints aiming at causing the server to perform the relatively
expensive
complaint validation. This is because it becomes blacklisted
by both the server and its peers the moment the invalid complaint is
ruled against it.
In addition, a malicious client cannot attack the server by sending
valid signed
messages with redundant valid complaints. Our protocol detects
duplicate complaints through
the use of time-stamps and caching of recent complaints.
Observation 4.8 A
malicious client
can always abandon any instance of the protocol.
In such case,
does not receive any credit, as
argued in Lemmas 4.1 to 4.3, even though
may have consumed
's resources.
This is a denial of service attack against
. Note that this attack would require that the malicious client
expends resources proportional to the resources of the victim
, therefore it is not particularly practical. Nevertheless, we prevent
blacklisted clients or clients that do not maintain paid accounts with
the content provider from launching such attack by having
issue a short-lived ticket
to authorized clients only.
is checked for validity by
(steps 4 and 6 in Section 3.4.2). In
addition,
may charge an authorized
for the issuance of tickets
effectively deterring
from maliciously expending both
's and the server's resources.
Owing to properties 4.1, 4.2, 4.3 and 4.5, and given that the content provider employs appropriate pricing schemes, Dandelion ensures that selfish (rational) clients increase their utility when they upload correct chunks and obtain virtual currency, while misbehaving clients cannot increase their utility. Consequently, Dandelion entices selfish clients to upload to their peers, resulting in a Nash equilibrium of cooperation.
This section describes a prototype C implementation of Dandelion that is suitable for cooperative content distribution of static content. It uses the openssl [5] library for cryptographic operations and standard file I/O system calls to efficiently manage credit information, which is stored in a simple file.
The server and the credit base are logical modules and could be distributed over a cluster to improve scalability. For simplicity, our current implementation combines the content provider and the credit base at a single server.
The server implementation is single-threaded and event-driven. The network operations are asynchronous, and data are transmitted over TCP.
Each client is assigned an entry in a credit file, which stores the credit as well as authentication and file access control information. Each entry has the same size and the client ID determines the offset of the entry of each client in the file, thus each entry can be efficiently accessed for both queries and updates.
The server queries and updates a client's credit from and to the credit file upon every transaction, Yet, it does not force commitment of the update to persistent storage. Instead, it relies on the OS to perform the commitment. If the server application crashes, the update will still be copied from the kernel buffer to persistent storage. Still, the OS may crash or the server may lose power before the updated data have been committed. However, in practice, a typical Dandelion deployment would run a stable operating system and use backup power supply. In addition, the server could mirror the credit base on multiple machines using high speed IP/Ethernet I/O. In addition, transactions would not involve very large amounts of money per user. Hence, we believe it is preferable not to incur the high cost of committing the credit updates to non-volatile memory after every transaction or after a few transactions (operations 12 and 13 in Table 1).
The client side is also single-threaded and event-driven. A client may leech or seed multiple files at a time. A client can be decomposed into two logical modules: a) the connection management module; and b) the peer-serving module.
The connection management module performs peering
and uploader discovery. With peering, each client
obtains a random partial
swarm view from the server and strives to connect to peers,
where
is the number of nodes in the
Dandelion swarm, as
communicated to the node by the server. As a result, the swarm
approximates a random graph with logarithmic out-degree, which
has been shown to have high connectivity[21]. With
uploader discovery, a client attempts to remain connected to a minimum
number of uploading peers. If the number of recent uploaders drops
below a threshold, a client requests from the server a new swarm view
and connects to the peers in the new view.
The peer-serving module performs content reconciliation and downloader selection. Content reconciliation refers to the client functionality for announcing recently received chunks, requesting missing chunks, requesting decryption keys for received encrypted chunks, and replying to chunk requests. Our implementation employs locally-rarest-random [39] scheduling in requesting missing chunks from clients. To efficiently utilize their downlink bandwidth using TCP, clients strive to at all times keep a specified number of outstanding chunk requests [26,40], which have been sent to a peer and have not been responded to.
With downloader selection, the system aims at making chunks available
to
the network as soon as possible. In the following description,
denotes the number of parallel downloaders. Dandelion's downloader
selection algorithm is similar to the seeder
choking algorithm used in the ``vanilla'' BitTorrent 4.0.2, as
documented in [41].
The algorithm is executed
by each client every 10 seconds. It is also executed when a when a peer
that is selected to be downloader disconnects. The algorithm proceeds
as follows: a) peers that are interested in the client's content are
ranked based on the time they were last selected to be downloaders
(most recent first); b) the client selects as downloaders the
top ranked peers; c) in case of a tie, the peer with the highest
download rate from the client is ranked higher; and d) the client
randomly selects an additional downloader from the non-selected nodes
that are interested in the client's content. Step (d) is performed in
expectation of discovering a fast downloader and to jumpstart peers
that recently joined the swarm.
This downloader selection algorithm aims at reducing the amount of
duplicate data a client needs to upload, before it has uploaded a full
copy of its content to the swarm. Downloader selection improves the
system's performance in the following additional ways. First, it limits
the number of peers a client concurrently uploads to, such that
complete chunks are made available to other peers and relayed by them
at faster rates. Second, given that all clients are connected to
roughly the same number of peers, it also limits the number of peers a
client concurrently downloads from to approximately
. As a result, the rate with which
the client downloads complete chunks increases. Last, by limiting the
number of connections over which clients upload, it avoids the
inefficiency and unfairness that is observed when many TCP flows share
a bottleneck link [46].
The number of peers that download from a client in parallel depends on the client's upload bandwidth. We have empirically determined that a good value for this parameter in the PlanetLab environment is 10.
The goal of our experimental evaluation is to demonstrate the viability and to identify the scalability limits of Dandelion's centralized and non-manipulable virtual-currency-based incentives.
We profile the cost of operations performed by the server aiming at identifying the performance bottlenecks of our design. The measurements are performed on a dual Pentium D 2.8GHZ/1MB CPU with 1GB RAM and 250GB/7200RPM HDD running Linux 2.6.5-1.358smp.
Table 1 lists the cost of per chunk Dandelion operations. In a flash crowd event, the main task of a Dandelion server is to: a) receive the decryption key request (operation 7); b) authenticate the request by computing an HMAC (operation 1); c) verify the ticket by computing an HMAC (operation 2); d) decrypt the encrypted decryption key (operation 3); e) query and update the credit of the two clients involved (operations 10 and 11); f) sign the decryption key response (operation 4); and g) send the decryption key response (operation 8).
As can be seen in the table, the per chunk cryptographic operations of the server (operations 1-4) are highly efficient (total 109 usec), as only symmetric cryptography is employed. The credit management operations (10 and 11) are also efficient (total 24 usec). On the other hand, the communication costs clearly dominate the processing costs, indicating that for 1Mb/s uplink and downlink, the downlink is the bottleneck.
The cost of a complaint is higher because in addition to verifying a ticket, it involves reading a chunk, encrypting it with the sender client's key (operation 5), and hashing the encrypted chunk (operation 6).
Note that the profiling of the server repeats the same operation multiple times. It does not consider the parallel processing of I/O and CPU operations. In addition, it does not include the cost of system calls and the cost of TCP/IP stack processing. Therefore, we refrain from deriving conclusions on the throughput of the server. Such conclusions are derived in the subsequent evaluation.
|
A Dandelion server mediates the chunk exchanges between its clients. The download throughput of clients in our system is bound by how fast a server can process their decryption key requests. Both the server's computational resources and bandwidth may become the performance bottleneck.
We deploy a Dandelion server that runs on the same machine
as the one used for Dandelion profiling. We also deploy
200 clients that run on distinct
PlanetLab hosts. The server machine shares a 100Mb/s Ethernet II link.
To mitigate bandwidth variability in the shared link and to emulate a
low cost server with an uplink and downlink that ranges from 1Mb/s to
5Mb/s, we rate-limit the server at the application layer.
In each experiment, the clients send requests for decryption keys to the server and we measure the aggregate rate with which all clients receive key responses. The server always queries and updates the credit base from and to the credit file without forcing commitment to disk. The specified per client request rate varies from 1 to 14 requests per second. For each request rate, the experiment duration was 10 minutes and the results were averaged over 10 runs. As the request rate increases and the server's receiver buffers become full, clients do not send new requests at the specified rate, due to TCP's flow control. When the request rate increases to the point that the server's resources become saturated, the key response rate from the server decreases.
Figure 3(a)
depicts the server's decryption
key throughput for various server bandwidth capacities. As the
bandwidth
increases from 1Mb/s to 3Mb/s, the server's decryption key response
throughput increases. This indicates that for 1Mb/s to 3Mb/s access
links, the bottleneck is the bandwidth. When the bandwidth limit is
4Mb/s and 5Mb/s, the server exhibits similar performance, which
suggests that the access link is
not saturated at 4Mb/s. The results show that a server running on our
commodity PC with 4Mb/s or 5Mb/s access link can process up to
1200 decryption key requests per second. This indicates that with a
128KB chunk size, this server may simultaneously support almost 1200
clients that download from each other
at 128KB/s. With a larger chunk size, each such client sends decryption
key requests at a slower rate, and the number of supported clients
increases.
Figures 3(b)
and 3(c)
show the average CPU and memory utilization at the server over the
duration of the above experiments. We observe that for 4Mb/s and 5Mb/s,
the server's CPU utilization reaches
100%, indicating that the bottleneck is the CPU. In Figure 3(c),
we see that the server consumes a very small portion of the available
memory.
The following experiments evaluate the performance of the Dandelion system on PlanetLab. We examine the impact of chunk size and the impact of seeding on the performance of the system. We also compare our system's performance to BitTorrent's. In all experiments we ran a Dandelion server within a PlanetLab VServer spawned on a highly available Xeon 3GHZ/2MB CPU and 2GB RAM machine. We rate-limit the server at 2Mb/s.
Leechers are given sufficient initial credit to completely download a file. Clients always respond to chunk requests from their selected downloaders and they never request chunks from the server. We do not rate-limit the Dandelion and BitTorrent clients, as a means for testing our system in heterogeneous Internet environments. To cover the bandwidth-delay product in PlanetLab, the TCP sender and receiver buffer size is set equal to 120KB.
For each configuration we repeat the experiment 10 times and we extract mean values and 95% confidence intervals over the swarm-wide mean file download completion times of each run. The file download completion time is the time that elapses between the moment the client contacts the server to start a download and the moment its download is completed.
This experiment aims at examining the tradeoffs involved in
selecting
the size of the chunk, the verifiable transaction unit in Dandelion.
Intuitively, since clients are able to serve a chunk only
as soon as they complete its download, a smaller chunk size yields a
more efficient distribution pipeline. In addition, when the file is
divided into many pieces, chunk scheduling techniques such as
rarest-first can be more effective, as there is sufficient content
entropy in
the network. Consequently, clients can promptly discover and download
content of interest. However, a smaller chunk size increases the rate
with which key requests are sent to the server, reducing the
scalability of the system. In addition, due to TCP's slow start, a
small chunk size cannot ensure high bandwidth utilization during the
TCP transfer
of any chunk.
In each configuration, we deploy around 40 Dandelion leechers and one initial seeder, i.e. a client that has the complete file before the distribution starts. Leechers start downloading files almost simultaneously. We deploy only 40 leechers to ensure that the server is not saturated, even if we use 64KB chunk size.
![]() |
Figure 4 shows the leecher mean download completion time as a function of the chunk size. For smaller files, e.g., the 10MB file, the system has the best performance for chunk size equal to 64KB. The system's performance degrades with the chunk size. As the file size increases, the beneficial impact of small chunks, becomes less significant. For example, for 250MB file, a 128KB chunk size yields notably better performance than a 64KB chunk size.
Based on the above results and further fine-tuning, in the rest of this evaluation, we use 128KB chunks.
One of Dandelion's main advantages is that it provides robust
incentives for clients to seed. We quantify the performance gains from
the existence of seeders in our system. In each experiment, we
deploy
200 leechers. Leechers start downloading the file almost
simultaneously, creating a flash crowd.
We show the impact of seeders by varying the probability that a
leecher remains online to seed a file after it completes its download.
In each experiment, a swarm has one initial seeder. Upon completion
of its download, each leecher stays in the swarm and seeds with
probability
. Probability
varies in 25% and 100%.
![]() |
Figure 5
depicts the mean download completion time over all
200 leechers as a function of the file size, for varying
. The results show the beneficial impact of seeders and the importance
of a mechanism to robustly incent seeding. For example, for a 250MB
file, we observe a swarm-wide mean download completion time of 674 sec
and 837 sec when leechers become seeders with 100% and 25% probability,
respectively. If we express the impact of seeders as the ratio of the
mean download
time for
over the mean
download time for
, we observe that the impact is reduced as the file size increases. The
larger the file is, the longer clients remain online to download it,
resulting in clients contributing their upload bandwidth for longer
periods. For smaller files however, leechers rely heavily on the
initial seeder and the leechers that become seeders to download their
content from. Therefore for small files, a reduction in probability
results in substantially longer download completion times.
Figure 5
also shows the
download completion times of
200 tit-for-tat compliant CTorrent 1.3.4 leechers. All BitTorrent
leechers remain online after their download completion (
).
The purpose of this illustration is to show that Dandelion can attain
performance comparable to the one achieved by BitTorrent,
although it employs a different downloader selection algorithm and
involves the server in each chunk exchange.
Although Dandelion appears to outperform BitTorrent for certain file sizes, we do not claim that it is in general a better-performing protocol. The performance of both protocols is highly dependent on numerous parameters, which we have not exhaustively analyzed.
We examine the distribution of credit during a Dandelion file distribution. The purpose of this measurement is to identify which type of clients tend to accumulate the most credit in swarms of similar configuration to ours.
Figure 6
shows the scatter plot of the
client's credit after a single 250MB file download by
200 leechers together
with the mean download rate of each client. In the experiment, there
is only one initial seeder. All nodes are given 100% of the credit
needed to download the file and they all become seeders upon download
completion. We observe that the seeder obtained the most credit during
the file distribution. This is expected, as a seeder is always in
position to upload useful content to its peers and our seeder had a
fast access link. Since fast downloaders obtain useful content
earlier in the distribution and are likely to have uplinks
proportional to their downlink, they should be able to deliver more
content and earn more credit. Our results confirm this intuition and
show that there is strong correlation between
average download rate and credit ratio, i.e. the product-moment
correlation coefficient is equal to 0.72.
We also observe that many clients uploaded substantially less than they downloaded. Indicatively, 25.8% of the clients had less than 70% of their initial credit.
![]() |
This paper describes Dandelion: a cooperative (P2P) system for the distribution of paid content. Dandelion's primary function is to enable a content provider to provide strong incentives for clients to contribute their uplink bandwidth.
Dandelion rewards selfish clients with virtual currency (credit) when they upload valid content to their peers and charges clients when they download content from selfish peers. Since Dandelion employs a non-manipulable cryptographic scheme for the fair exchange of content uploads for credit, the content provider is able to redeem a client's credit for monetary rewards. Thus, it provides strong incentives for clients to seed content and to not free-ride.
Our experimental results show that seeding substantially improves swarm-wide performance. They also show that a Dandelion server running on commodity hardware and with moderate bandwidth can scale to a few thousand clients. Dandelion's deployment in medium size swarms demonstrates that it can attain performance that is comparable to BitTorrent. These facts demonstrate the plausibility of our design choice: centralizing the incentive mechanism in order to increase resource availability in P2P content distribution networks.
We are thankful to Nikitas Liogkas, Eddie Kohler and the anonymous reviewers for their extensive and fruitful feedback. We also thank Rex Chen, Dehn Sy and Lichun Bao with Calit2 for providing space and equipment for our experiments. This work was supported in part by NSF award CNS-0627166.
Last changed: 30 May 2007 ch |