mirror of
https://github.com/tahoe-lafs/tahoe-lafs.git
synced 2024-12-24 07:06:41 +00:00
a few documentation and naming convention updates
Notable: the argument to make REPORTER has been renamed to TRIALARGS.
This commit is contained in:
parent
a7f954ac06
commit
1ac09840a4
4
Makefile
4
Makefile
@ -159,12 +159,12 @@ check-pyopenssl-dep:
|
||||
# too.
|
||||
TEST=allmydata
|
||||
|
||||
# use 'make test REPORTER=--reporter=bwverbose' from buildbot, to
|
||||
# use 'make test TRIALARGS=--reporter=bwverbose' from buildbot, to
|
||||
# suppress the ansi color sequences
|
||||
|
||||
test: .built .checked-deps
|
||||
$(PP) \
|
||||
$(TRIAL) $(REPORTER) $(TEST)
|
||||
$(TRIAL) $(TRIALARGS) $(TEST)
|
||||
|
||||
test-figleaf: .built .checked-deps
|
||||
rm -f .figleaf
|
||||
|
@ -114,7 +114,7 @@ class Node:
|
||||
self.introducer.delete(fileid)
|
||||
return True
|
||||
|
||||
class IntroducerAndVdrive:
|
||||
class Introducer:
|
||||
def __init__(self, simulator):
|
||||
self.living_files = {}
|
||||
self.utilization = 0 # total size of all active files
|
||||
@ -169,7 +169,7 @@ class Simulator:
|
||||
self.rrd = RRD("/tmp/utilization.rrd", ds=[ds], rra=[rra], start=self.time)
|
||||
self.rrd.create()
|
||||
|
||||
self.introducer = q = IntroducerAndVdrive(self)
|
||||
self.introducer = q = Introducer(self)
|
||||
self.all_nodes = [Node(randomid(), q, self)
|
||||
for i in range(self.NUM_NODES)]
|
||||
q.all_nodes = self.all_nodes
|
||||
|
@ -17,16 +17,16 @@ The goal of the encoder is to turn the original file into a series of
|
||||
is a different host, but for small grids there may be overlap). The number
|
||||
of shares is chosen to hit our reliability goals (more shares on more
|
||||
machines means more reliability), and is limited by overhead (proportional to
|
||||
numshares or log(numshares)) and the encoding technology in use (Reed-Solomon
|
||||
only permits 256 shares total). It is also constrained by the amount of data
|
||||
numshares or log(numshares)) and the encoding technology in use (zfec permits
|
||||
only 256 shares total). It is also constrained by the amount of data
|
||||
we want to send to each host. For estimating purposes, think of 10 shares
|
||||
out of which we need 3 to reconstruct the file.
|
||||
|
||||
The encoder starts by cutting the original file into segments. All segments
|
||||
except the last are of equal size. The segment size is chosen to constrain
|
||||
the memory footprint (which will probably vary between 1x and 4x segment
|
||||
size) and to constrain the overhead (which will be proportional to either the
|
||||
number of segments or log(number of segments)).
|
||||
size) and to constrain the overhead (which will be proportional to
|
||||
log(number of segments)).
|
||||
|
||||
|
||||
Each segment (A,B,C) is read into memory, encrypted, and encoded into
|
||||
|
@ -1,4 +1,3 @@
|
||||
|
||||
import re, os.path, urllib
|
||||
import simplejson
|
||||
from twisted.application import service
|
||||
|
@ -39,7 +39,7 @@ try:
|
||||
try:
|
||||
current = resource.getrlimit(resource.RLIMIT_NOFILE)
|
||||
except AttributeError:
|
||||
# we're probably missing RLIMIT_NOFILE, maybe this is windows
|
||||
# we're probably missing RLIMIT_NOFILE
|
||||
return
|
||||
|
||||
if current[0] >= 1024:
|
||||
|
Loading…
Reference in New Issue
Block a user