diff --git a/integration/util.py b/integration/util.py index a6543c34d..02cda655d 100644 --- a/integration/util.py +++ b/integration/util.py @@ -46,7 +46,7 @@ from allmydata.util.configutil import ( write_config, ) from allmydata import client -from allmydata.interfaces import DEFAULT_MAX_SEGMENT_SIZE +from allmydata.interfaces import DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE import pytest_twisted @@ -774,7 +774,7 @@ async def reconfigure(reactor, request, node: TahoeProcess, config.write_private_config("convergence", base32.b2a(convergence)) if max_segment_size is not None: - cur_segment_size = int(config.get_config("client", "shares._max_immutable_segment_size_for_testing", DEFAULT_MAX_SEGMENT_SIZE)) + cur_segment_size = int(config.get_config("client", "shares._max_immutable_segment_size_for_testing", DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE)) if cur_segment_size != max_segment_size: changed = True config.set_config( diff --git a/misc/simulators/storage-overhead.py b/misc/simulators/storage-overhead.py index 5a741834e..9959f5575 100644 --- a/misc/simulators/storage-overhead.py +++ b/misc/simulators/storage-overhead.py @@ -5,7 +5,7 @@ from __future__ import print_function import sys, math from allmydata import uri, storage from allmydata.immutable import upload -from allmydata.interfaces import DEFAULT_MAX_SEGMENT_SIZE +from allmydata.interfaces import DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE from allmydata.util import mathutil def roundup(size, blocksize=4096): @@ -26,7 +26,7 @@ class BigFakeString(object): def tell(self): return self.fp -def calc(filesize, params=(3,7,10), segsize=DEFAULT_MAX_SEGMENT_SIZE): +def calc(filesize, params=(3,7,10), segsize=DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE): num_shares = params[2] if filesize <= upload.Uploader.URI_LIT_SIZE_THRESHOLD: urisize = len(uri.LiteralFileURI("A"*filesize).to_string()) diff --git a/src/allmydata/client.py b/src/allmydata/client.py index c06961c8c..dfd1c3c81 100644 --- a/src/allmydata/client.py +++ b/src/allmydata/client.py @@ -50,7 +50,7 @@ from allmydata.interfaces import ( IStatsProducer, SDMF_VERSION, MDMF_VERSION, - DEFAULT_MAX_SEGMENT_SIZE, + DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE, IFoolscapStoragePlugin, IAnnounceableStorageServer, ) @@ -607,7 +607,7 @@ class _Client(node.Node, pollmixin.PollMixin): DEFAULT_ENCODING_PARAMETERS = {"k": 3, "happy": 7, "n": 10, - "max_segment_size": DEFAULT_MAX_SEGMENT_SIZE, + "max_segment_size": DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE, } def __init__(self, config, main_tub, i2p_provider, tor_provider, introducer_clients, diff --git a/src/allmydata/immutable/downloader/node.py b/src/allmydata/immutable/downloader/node.py index ec4f751df..a1ef4b485 100644 --- a/src/allmydata/immutable/downloader/node.py +++ b/src/allmydata/immutable/downloader/node.py @@ -19,7 +19,7 @@ from foolscap.api import eventually from allmydata import uri from allmydata.codec import CRSDecoder from allmydata.util import base32, log, hashutil, mathutil, observer -from allmydata.interfaces import DEFAULT_MAX_SEGMENT_SIZE +from allmydata.interfaces import DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE from allmydata.hashtree import IncompleteHashTree, BadHashError, \ NotEnoughHashesError @@ -49,7 +49,7 @@ class DownloadNode(object): """Internal class which manages downloads and holds state. External callers use CiphertextFileNode instead.""" - default_max_segment_size = DEFAULT_MAX_SEGMENT_SIZE + default_max_segment_size = DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE # Share._node points to me def __init__(self, verifycap, storage_broker, secret_holder, diff --git a/src/allmydata/immutable/upload.py b/src/allmydata/immutable/upload.py index 6b9b48f6a..06482688f 100644 --- a/src/allmydata/immutable/upload.py +++ b/src/allmydata/immutable/upload.py @@ -48,7 +48,7 @@ from allmydata.util.rrefutil import add_version_to_remote_reference from allmydata.interfaces import IUploadable, IUploader, IUploadResults, \ IEncryptedUploadable, RIEncryptedUploadable, IUploadStatus, \ NoServersError, InsufficientVersionError, UploadUnhappinessError, \ - DEFAULT_MAX_SEGMENT_SIZE, IPeerSelector + DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE, IPeerSelector from allmydata.immutable import layout from io import BytesIO @@ -1692,7 +1692,7 @@ class AssistedUploader(object): class BaseUploadable(object): # this is overridden by max_segment_size - default_max_segment_size = DEFAULT_MAX_SEGMENT_SIZE + default_max_segment_size = DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE default_params_set = False max_segment_size = None diff --git a/src/allmydata/interfaces.py b/src/allmydata/interfaces.py index 8f6a0c37a..493a4a31f 100644 --- a/src/allmydata/interfaces.py +++ b/src/allmydata/interfaces.py @@ -41,7 +41,7 @@ URI = StringConstraint(300) # kind of arbitrary MAX_BUCKETS = 256 # per peer -- zfec offers at most 256 shares per file -DEFAULT_MAX_SEGMENT_SIZE = 1024*1024 +DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE = 1024*1024 ShareData = StringConstraint(None) URIExtensionData = StringConstraint(1000) diff --git a/src/allmydata/mutable/publish.py b/src/allmydata/mutable/publish.py index a7bca6cba..5f689dbc0 100644 --- a/src/allmydata/mutable/publish.py +++ b/src/allmydata/mutable/publish.py @@ -36,7 +36,7 @@ from allmydata.mutable.layout import get_version_from_checkstring,\ SDMFSlotWriteProxy KiB = 1024 -DEFAULT_MAX_SEGMENT_SIZE = 128 * KiB +DEFAULT_MUTABLE_MAX_SEGMENT_SIZE = 128 * KiB PUSHING_BLOCKS_STATE = 0 PUSHING_EVERYTHING_ELSE_STATE = 1 DONE_STATE = 2 @@ -367,7 +367,7 @@ class Publish(object): self.data = newdata self.datalength = newdata.get_size() - #if self.datalength >= DEFAULT_MAX_SEGMENT_SIZE: + #if self.datalength >= DEFAULT_MUTABLE_MAX_SEGMENT_SIZE: # self._version = MDMF_VERSION #else: # self._version = SDMF_VERSION @@ -551,7 +551,7 @@ class Publish(object): def setup_encoding_parameters(self, offset=0): if self._version == MDMF_VERSION: - segment_size = DEFAULT_MAX_SEGMENT_SIZE # 128 KiB by default + segment_size = DEFAULT_MUTABLE_MAX_SEGMENT_SIZE # 128 KiB by default else: segment_size = self.datalength # SDMF is only one segment # this must be a multiple of self.required_shares diff --git a/src/allmydata/test/mutable/test_update.py b/src/allmydata/test/mutable/test_update.py index c3ba1e9f7..1c91590bd 100644 --- a/src/allmydata/test/mutable/test_update.py +++ b/src/allmydata/test/mutable/test_update.py @@ -20,7 +20,7 @@ from testtools.matchers import ( from twisted.internet import defer from allmydata.interfaces import MDMF_VERSION from allmydata.mutable.filenode import MutableFileNode -from allmydata.mutable.publish import MutableData, DEFAULT_MAX_SEGMENT_SIZE +from allmydata.mutable.publish import MutableData, DEFAULT_MUTABLE_MAX_SEGMENT_SIZE from ..no_network import GridTestMixin from .. import common_util as testutil @@ -180,7 +180,7 @@ class Update(GridTestMixin, AsyncTestCase, testutil.ShouldFailMixin): # long -- this is 7 segments in the default segment size. So we # need to add 2 segments worth of data to push it over a # power-of-two boundary. - segment = b"a" * DEFAULT_MAX_SEGMENT_SIZE + segment = b"a" * DEFAULT_MUTABLE_MAX_SEGMENT_SIZE new_data = self.data + (segment * 2) d0 = self.do_upload_mdmf() def _run(ign): @@ -232,9 +232,9 @@ class Update(GridTestMixin, AsyncTestCase, testutil.ShouldFailMixin): return d0 def test_multiple_segment_replace(self): - replace_offset = 2 * DEFAULT_MAX_SEGMENT_SIZE + replace_offset = 2 * DEFAULT_MUTABLE_MAX_SEGMENT_SIZE new_data = self.data[:replace_offset] - new_segment = b"a" * DEFAULT_MAX_SEGMENT_SIZE + new_segment = b"a" * DEFAULT_MUTABLE_MAX_SEGMENT_SIZE new_data += 2 * new_segment new_data += b"replaced" rest_offset = len(new_data) diff --git a/src/allmydata/test/test_system.py b/src/allmydata/test/test_system.py index 9849c6b30..1b8a8dbfe 100644 --- a/src/allmydata/test/test_system.py +++ b/src/allmydata/test/test_system.py @@ -37,7 +37,7 @@ from allmydata.util.consumer import MemoryConsumer, download_to_data from allmydata.util.deferredutil import async_to_deferred from allmydata.interfaces import IDirectoryNode, IFileNode, \ NoSuchChildError, NoSharesError, SDMF_VERSION, MDMF_VERSION, \ - DEFAULT_MAX_SEGMENT_SIZE + DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE from allmydata.monitor import Monitor from allmydata.mutable.common import NotWriteableError from allmydata.mutable import layout as mutable_layout @@ -1846,7 +1846,7 @@ class SystemTest(SystemTestMixin, RunBinTahoeMixin, unittest.TestCase): uploadable.max_segment_size = upload_segment_size results = await uploader.upload(uploadable) - assert DownloadNode.default_max_segment_size == DEFAULT_MAX_SEGMENT_SIZE + assert DownloadNode.default_max_segment_size == DEFAULT_IMMUTABLE_MAX_SEGMENT_SIZE self.patch(DownloadNode, "default_max_segment_size", download_segment_size) uri = results.get_uri() node = self.clients[1].create_node_from_uri(uri)