mirror of
https://github.com/tahoe-lafs/tahoe-lafs.git
synced 2024-12-29 01:08:54 +00:00
3d771132a8
Complete the getter-based transformation, by hiding ".uri" and updating callers to use get_uri(). Also don't set a dummy self._uri, leave it undefined until someone calls set_uri().
335 lines
15 KiB
Python
335 lines
15 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
import os, shutil
|
|
from twisted.trial import unittest
|
|
from twisted.internet import defer
|
|
from allmydata import uri
|
|
from allmydata.util.consumer import download_to_data
|
|
from allmydata.immutable import upload
|
|
from allmydata.mutable.common import UnrecoverableFileError
|
|
from allmydata.mutable.publish import MutableData
|
|
from allmydata.storage.common import storage_index_to_dir
|
|
from allmydata.test.no_network import GridTestMixin
|
|
from allmydata.test.common import ShouldFailMixin
|
|
from allmydata.util.pollmixin import PollMixin
|
|
from allmydata.interfaces import NotEnoughSharesError
|
|
|
|
immutable_plaintext = "data" * 10000
|
|
mutable_plaintext = "muta" * 10000
|
|
|
|
class HungServerDownloadTest(GridTestMixin, ShouldFailMixin, PollMixin,
|
|
unittest.TestCase):
|
|
# Many of these tests take around 60 seconds on François's ARM buildslave:
|
|
# http://tahoe-lafs.org/buildbot/builders/FranXois%20lenny-armv5tel
|
|
# allmydata.test.test_hung_server.HungServerDownloadTest.test_2_good_8_broken_duplicate_share_fail
|
|
# once ERRORed after 197 seconds on Midnight Magic's NetBSD buildslave:
|
|
# http://tahoe-lafs.org/buildbot/builders/MM%20netbsd4%20i386%20warp
|
|
# MM's buildslave varies a lot in how long it takes to run tests.
|
|
|
|
timeout = 240
|
|
|
|
def _break(self, servers):
|
|
for (id, ss) in servers:
|
|
self.g.break_server(id)
|
|
|
|
def _hang(self, servers, **kwargs):
|
|
for (id, ss) in servers:
|
|
self.g.hang_server(id, **kwargs)
|
|
|
|
def _unhang(self, servers, **kwargs):
|
|
for (id, ss) in servers:
|
|
self.g.unhang_server(id, **kwargs)
|
|
|
|
def _hang_shares(self, shnums, **kwargs):
|
|
# hang all servers who are holding the given shares
|
|
hung_serverids = set()
|
|
for (i_shnum, i_serverid, i_sharefile) in self.shares:
|
|
if i_shnum in shnums:
|
|
if i_serverid not in hung_serverids:
|
|
self.g.hang_server(i_serverid, **kwargs)
|
|
hung_serverids.add(i_serverid)
|
|
|
|
def _delete_all_shares_from(self, servers):
|
|
serverids = [id for (id, ss) in servers]
|
|
for (i_shnum, i_serverid, i_sharefile) in self.shares:
|
|
if i_serverid in serverids:
|
|
os.unlink(i_sharefile)
|
|
|
|
def _corrupt_all_shares_in(self, servers, corruptor_func):
|
|
serverids = [id for (id, ss) in servers]
|
|
for (i_shnum, i_serverid, i_sharefile) in self.shares:
|
|
if i_serverid in serverids:
|
|
self._corrupt_share((i_shnum, i_sharefile), corruptor_func)
|
|
|
|
def _copy_all_shares_from(self, from_servers, to_server):
|
|
serverids = [id for (id, ss) in from_servers]
|
|
for (i_shnum, i_serverid, i_sharefile) in self.shares:
|
|
if i_serverid in serverids:
|
|
self._copy_share((i_shnum, i_sharefile), to_server)
|
|
|
|
def _copy_share(self, share, to_server):
|
|
(sharenum, sharefile) = share
|
|
(id, ss) = to_server
|
|
shares_dir = os.path.join(ss.original.storedir, "shares")
|
|
si = uri.from_string(self.uri).get_storage_index()
|
|
si_dir = os.path.join(shares_dir, storage_index_to_dir(si))
|
|
if not os.path.exists(si_dir):
|
|
os.makedirs(si_dir)
|
|
new_sharefile = os.path.join(si_dir, str(sharenum))
|
|
shutil.copy(sharefile, new_sharefile)
|
|
self.shares = self.find_uri_shares(self.uri)
|
|
# Make sure that the storage server has the share.
|
|
self.failUnless((sharenum, ss.original.my_nodeid, new_sharefile)
|
|
in self.shares)
|
|
|
|
def _corrupt_share(self, share, corruptor_func):
|
|
(sharenum, sharefile) = share
|
|
data = open(sharefile, "rb").read()
|
|
newdata = corruptor_func(data)
|
|
os.unlink(sharefile)
|
|
wf = open(sharefile, "wb")
|
|
wf.write(newdata)
|
|
wf.close()
|
|
|
|
def _set_up(self, mutable, testdir, num_clients=1, num_servers=10):
|
|
self.mutable = mutable
|
|
if mutable:
|
|
self.basedir = "hung_server/mutable_" + testdir
|
|
else:
|
|
self.basedir = "hung_server/immutable_" + testdir
|
|
|
|
self.set_up_grid(num_clients=num_clients, num_servers=num_servers)
|
|
|
|
self.c0 = self.g.clients[0]
|
|
nm = self.c0.nodemaker
|
|
self.servers = sorted([(s.get_serverid(), s.get_rref())
|
|
for s in nm.storage_broker.get_connected_servers()])
|
|
self.servers = self.servers[5:] + self.servers[:5]
|
|
|
|
if mutable:
|
|
uploadable = MutableData(mutable_plaintext)
|
|
d = nm.create_mutable_file(uploadable)
|
|
def _uploaded_mutable(node):
|
|
self.uri = node.get_uri()
|
|
self.shares = self.find_uri_shares(self.uri)
|
|
d.addCallback(_uploaded_mutable)
|
|
else:
|
|
data = upload.Data(immutable_plaintext, convergence="")
|
|
d = self.c0.upload(data)
|
|
def _uploaded_immutable(upload_res):
|
|
self.uri = upload_res.get_uri()
|
|
self.shares = self.find_uri_shares(self.uri)
|
|
d.addCallback(_uploaded_immutable)
|
|
return d
|
|
|
|
def _start_download(self):
|
|
n = self.c0.create_node_from_uri(self.uri)
|
|
if self.mutable:
|
|
d = n.download_best_version()
|
|
else:
|
|
d = download_to_data(n)
|
|
return d
|
|
|
|
def _wait_for_data(self, n):
|
|
if self.mutable:
|
|
d = n.download_best_version()
|
|
else:
|
|
d = download_to_data(n)
|
|
return d
|
|
|
|
def _check(self, resultingdata):
|
|
if self.mutable:
|
|
self.failUnlessEqual(resultingdata, mutable_plaintext)
|
|
else:
|
|
self.failUnlessEqual(resultingdata, immutable_plaintext)
|
|
|
|
def _download_and_check(self):
|
|
d = self._start_download()
|
|
d.addCallback(self._check)
|
|
return d
|
|
|
|
def _should_fail_download(self):
|
|
if self.mutable:
|
|
return self.shouldFail(UnrecoverableFileError, self.basedir,
|
|
"no recoverable versions",
|
|
self._download_and_check)
|
|
else:
|
|
return self.shouldFail(NotEnoughSharesError, self.basedir,
|
|
"ran out of shares",
|
|
self._download_and_check)
|
|
|
|
|
|
def test_10_good_sanity_check(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_10_good_sanity_check"))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_10_good_copied_share(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_10_good_copied_share"))
|
|
d.addCallback(lambda ign: self._copy_all_shares_from(self.servers[2:3], self.servers[0]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_3_good_7_noshares(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_3_good_7_noshares"))
|
|
d.addCallback(lambda ign: self._delete_all_shares_from(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_2_good_8_broken_fail(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_2_good_8_broken_fail"))
|
|
d.addCallback(lambda ign: self._break(self.servers[2:]))
|
|
d.addCallback(lambda ign: self._should_fail_download())
|
|
return d
|
|
|
|
def test_2_good_8_noshares_fail(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_2_good_8_noshares_fail"))
|
|
d.addCallback(lambda ign: self._delete_all_shares_from(self.servers[2:]))
|
|
d.addCallback(lambda ign: self._should_fail_download())
|
|
return d
|
|
|
|
def test_2_good_8_broken_copied_share(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_2_good_8_broken_copied_share"))
|
|
d.addCallback(lambda ign: self._copy_all_shares_from(self.servers[2:3], self.servers[0]))
|
|
d.addCallback(lambda ign: self._break(self.servers[2:]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_2_good_8_broken_duplicate_share_fail(self):
|
|
d = defer.succeed(None)
|
|
for mutable in [False, True]:
|
|
d.addCallback(lambda ign, mutable=mutable: self._set_up(mutable, "test_2_good_8_broken_duplicate_share_fail"))
|
|
d.addCallback(lambda ign: self._copy_all_shares_from(self.servers[1:2], self.servers[0]))
|
|
d.addCallback(lambda ign: self._break(self.servers[2:]))
|
|
d.addCallback(lambda ign: self._should_fail_download())
|
|
return d
|
|
|
|
def test_3_good_7_hung_immutable(self):
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(False, "test_3_good_7_hung"))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_5_overdue_immutable(self):
|
|
# restrict the ShareFinder to only allow 5 outstanding requests, and
|
|
# arrange for the first 5 servers to hang. Then trigger the OVERDUE
|
|
# timers (simulating 10 seconds passed), at which point the
|
|
# ShareFinder should send additional queries and finish the download
|
|
# quickly. If we didn't have OVERDUE timers, this test would fail by
|
|
# timing out.
|
|
done = []
|
|
d = self._set_up(False, "test_5_overdue_immutable")
|
|
def _reduce_max_outstanding_requests_and_download(ign):
|
|
self._hang_shares(range(5))
|
|
n = self.c0.create_node_from_uri(self.uri)
|
|
n._cnode._maybe_create_download_node()
|
|
self._sf = n._cnode._node._sharefinder
|
|
self._sf.max_outstanding_requests = 5
|
|
self._sf.OVERDUE_TIMEOUT = 1000.0
|
|
d2 = download_to_data(n)
|
|
# start download, but don't wait for it to complete yet
|
|
def _done(res):
|
|
done.append(res) # we will poll for this later
|
|
d2.addBoth(_done)
|
|
d.addCallback(_reduce_max_outstanding_requests_and_download)
|
|
from foolscap.eventual import fireEventually, flushEventualQueue
|
|
# wait here a while
|
|
d.addCallback(lambda res: fireEventually(res))
|
|
d.addCallback(lambda res: flushEventualQueue())
|
|
d.addCallback(lambda ign: self.failIf(done))
|
|
def _check_waiting(ign):
|
|
# all the share requests should now be stuck waiting
|
|
self.failUnlessEqual(len(self._sf.pending_requests), 5)
|
|
# but none should be marked as OVERDUE until the timers expire
|
|
self.failUnlessEqual(len(self._sf.overdue_requests), 0)
|
|
d.addCallback(_check_waiting)
|
|
def _mark_overdue(ign):
|
|
# declare four requests overdue, allowing new requests to take
|
|
# their place, and leaving one stuck. The finder will keep
|
|
# sending requests until there are 5 non-overdue ones
|
|
# outstanding, at which point we'll have 4 OVERDUE, 1
|
|
# stuck-but-not-overdue, and 4 live requests. All 4 live requests
|
|
# will retire before the download is complete and the ShareFinder
|
|
# is shut off. That will leave 4 OVERDUE and 1
|
|
# stuck-but-not-overdue, for a total of 5 requests in in
|
|
# _sf.pending_requests
|
|
for t in self._sf.overdue_timers.values()[:4]:
|
|
t.reset(-1.0)
|
|
# the timers ought to fire before the eventual-send does
|
|
return fireEventually()
|
|
d.addCallback(_mark_overdue)
|
|
def _we_are_done():
|
|
return bool(done)
|
|
d.addCallback(lambda ign: self.poll(_we_are_done))
|
|
def _check_done(ign):
|
|
self.failUnlessEqual(done, [immutable_plaintext])
|
|
self.failUnlessEqual(len(self._sf.pending_requests), 5)
|
|
self.failUnlessEqual(len(self._sf.overdue_requests), 4)
|
|
d.addCallback(_check_done)
|
|
return d
|
|
|
|
def test_2_good_8_hung_then_1_recovers_immutable(self):
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(False, "test_2_good_8_hung_then_1_recovers"))
|
|
d.addCallback(lambda ign: self._hang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._unhang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_2_good_8_hung_then_1_recovers_with_2_shares_immutable(self):
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(False, "test_2_good_8_hung_then_1_recovers_with_2_shares"))
|
|
d.addCallback(lambda ign: self._copy_all_shares_from(self.servers[0:1], self.servers[2]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._unhang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
# The tests below do not currently pass for mutable files. The
|
|
# mutable-file downloader does not yet handle hung servers, and the tests
|
|
# hang forever (hence the use of SkipTest rather than .todo)
|
|
|
|
def test_3_good_7_hung_mutable(self):
|
|
raise unittest.SkipTest("still broken")
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(True, "test_3_good_7_hung"))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_2_good_8_hung_then_1_recovers_mutable(self):
|
|
raise unittest.SkipTest("still broken")
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(True, "test_2_good_8_hung_then_1_recovers"))
|
|
d.addCallback(lambda ign: self._hang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._unhang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|
|
|
|
def test_2_good_8_hung_then_1_recovers_with_2_shares_mutable(self):
|
|
raise unittest.SkipTest("still broken")
|
|
d = defer.succeed(None)
|
|
d.addCallback(lambda ign: self._set_up(True, "test_2_good_8_hung_then_1_recovers_with_2_shares"))
|
|
d.addCallback(lambda ign: self._copy_all_shares_from(self.servers[0:1], self.servers[2]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._hang(self.servers[3:]))
|
|
d.addCallback(lambda ign: self._unhang(self.servers[2:3]))
|
|
d.addCallback(lambda ign: self._download_and_check())
|
|
return d
|