mirror of
https://github.com/tahoe-lafs/tahoe-lafs.git
synced 2024-12-24 07:06:41 +00:00
switch from base62 to base32 for storage indices, switch from z-base-32 to rfc 3548 base-32 for everything, separate out base32 encoding from idlib
This commit is contained in:
parent
481f3649ec
commit
7c6de95bc6
@ -15,7 +15,7 @@ from allmydata.checker import Checker
|
||||
from allmydata.offloaded import Helper
|
||||
from allmydata.control import ControlServer
|
||||
from allmydata.introducer import IntroducerClient
|
||||
from allmydata.util import hashutil, idlib, testutil
|
||||
from allmydata.util import hashutil, base32, testutil
|
||||
from allmydata.filenode import FileNode
|
||||
from allmydata.dirnode import NewDirectoryNode
|
||||
from allmydata.mutable import MutableFileNode
|
||||
@ -102,9 +102,9 @@ class Client(node.Node, testutil.PollMixin):
|
||||
|
||||
def init_lease_secret(self):
|
||||
def make_secret():
|
||||
return idlib.b2a(os.urandom(hashutil.CRYPTO_VAL_SIZE)) + "\n"
|
||||
return base32.b2a(os.urandom(hashutil.CRYPTO_VAL_SIZE)) + "\n"
|
||||
secret_s = self.get_or_create_private_config("secret", make_secret)
|
||||
self._lease_secret = idlib.a2b(secret_s)
|
||||
self._lease_secret = base32.a2b(secret_s)
|
||||
|
||||
def init_storage(self):
|
||||
# should we run a storage server (and publish it for others to use)?
|
||||
|
@ -6,7 +6,7 @@ from twisted.internet.interfaces import IPushProducer, IConsumer
|
||||
from twisted.application import service
|
||||
from foolscap.eventual import eventually
|
||||
|
||||
from allmydata.util import idlib, mathutil, hashutil, log
|
||||
from allmydata.util import base32, mathutil, hashutil, log
|
||||
from allmydata.util.assertutil import _assert
|
||||
from allmydata import codec, hashtree, storage, uri
|
||||
from allmydata.interfaces import IDownloadTarget, IDownloader, IFileURI, \
|
||||
@ -70,7 +70,7 @@ class Output:
|
||||
crypttext_leaves = {self._segment_number: ch.digest()}
|
||||
self.log(format="crypttext leaf hash (%(bytes)sB) [%(segnum)d] is %(hash)s",
|
||||
bytes=len(crypttext),
|
||||
segnum=self._segment_number, hash=idlib.b2a(ch.digest()),
|
||||
segnum=self._segment_number, hash=base32.b2a(ch.digest()),
|
||||
level=log.NOISY)
|
||||
self._crypttext_hash_tree.set_hashes(leaves=crypttext_leaves)
|
||||
|
||||
@ -86,7 +86,7 @@ class Output:
|
||||
plaintext_leaves = {self._segment_number: ph.digest()}
|
||||
self.log(format="plaintext leaf hash (%(bytes)sB) [%(segnum)d] is %(hash)s",
|
||||
bytes=len(plaintext),
|
||||
segnum=self._segment_number, hash=idlib.b2a(ph.digest()),
|
||||
segnum=self._segment_number, hash=base32.b2a(ph.digest()),
|
||||
level=log.NOISY)
|
||||
self._plaintext_hash_tree.set_hashes(leaves=plaintext_leaves)
|
||||
|
||||
@ -180,7 +180,7 @@ class ValidatedBucket:
|
||||
#log.msg("checking block_hash(shareid=%d, blocknum=%d) len=%d "
|
||||
# "%r .. %r: %s" %
|
||||
# (self.sharenum, blocknum, len(blockdata),
|
||||
# blockdata[:50], blockdata[-50:], idlib.b2a(blockhash)))
|
||||
# blockdata[:50], blockdata[-50:], base32.b2a(blockhash)))
|
||||
|
||||
# we always validate the blockhash
|
||||
bh = dict(enumerate(blockhashes))
|
||||
@ -203,22 +203,22 @@ class ValidatedBucket:
|
||||
received from the remote peer were bad.""")
|
||||
log.msg(" have self._share_hash: %s" % bool(self._share_hash))
|
||||
log.msg(" block length: %d" % len(blockdata))
|
||||
log.msg(" block hash: %s" % idlib.b2a_or_none(blockhash))
|
||||
log.msg(" block hash: %s" % base32.b2a_or_none(blockhash))
|
||||
if len(blockdata) < 100:
|
||||
log.msg(" block data: %r" % (blockdata,))
|
||||
else:
|
||||
log.msg(" block data start/end: %r .. %r" %
|
||||
(blockdata[:50], blockdata[-50:]))
|
||||
log.msg(" root hash: %s" % idlib.b2a(self._roothash))
|
||||
log.msg(" root hash: %s" % base32.b2a(self._roothash))
|
||||
log.msg(" share hash tree:\n" + self.share_hash_tree.dump())
|
||||
log.msg(" block hash tree:\n" + self.block_hash_tree.dump())
|
||||
lines = []
|
||||
for i,h in sorted(sharehashes):
|
||||
lines.append("%3d: %s" % (i, idlib.b2a_or_none(h)))
|
||||
lines.append("%3d: %s" % (i, base32.b2a_or_none(h)))
|
||||
log.msg(" sharehashes:\n" + "\n".join(lines) + "\n")
|
||||
lines = []
|
||||
for i,h in enumerate(blockhashes):
|
||||
lines.append("%3d: %s" % (i, idlib.b2a_or_none(h)))
|
||||
lines.append("%3d: %s" % (i, base32.b2a_or_none(h)))
|
||||
log.msg(" blockhashes:\n" + "\n".join(lines) + "\n")
|
||||
raise
|
||||
|
||||
@ -782,13 +782,13 @@ class FileDownloader:
|
||||
if self.check_crypttext_hash:
|
||||
_assert(self._crypttext_hash == self._output.crypttext_hash,
|
||||
"bad crypttext_hash: computed=%s, expected=%s" %
|
||||
(idlib.b2a(self._output.crypttext_hash),
|
||||
idlib.b2a(self._crypttext_hash)))
|
||||
(base32.b2a(self._output.crypttext_hash),
|
||||
base32.b2a(self._crypttext_hash)))
|
||||
if self.check_plaintext_hash:
|
||||
_assert(self._plaintext_hash == self._output.plaintext_hash,
|
||||
"bad plaintext_hash: computed=%s, expected=%s" %
|
||||
(idlib.b2a(self._output.plaintext_hash),
|
||||
idlib.b2a(self._plaintext_hash)))
|
||||
(base32.b2a(self._output.plaintext_hash),
|
||||
base32.b2a(self._plaintext_hash)))
|
||||
_assert(self._output.length == self._size,
|
||||
got=self._output.length, expected=self._size)
|
||||
return self._output.finish()
|
||||
|
@ -6,7 +6,7 @@ from twisted.internet import defer
|
||||
from foolscap import eventual
|
||||
from allmydata import storage, uri
|
||||
from allmydata.hashtree import HashTree
|
||||
from allmydata.util import mathutil, hashutil, idlib, log
|
||||
from allmydata.util import mathutil, hashutil, base32, log
|
||||
from allmydata.util.assertutil import _assert, precondition
|
||||
from allmydata.codec import CRSEncoder
|
||||
from allmydata.interfaces import IEncoder, IStorageBucketWriter, \
|
||||
@ -435,11 +435,11 @@ class Encoder(object):
|
||||
d = self.send_subshare(shareid, segnum, subshare, lognum)
|
||||
dl.append(d)
|
||||
subshare_hash = hashutil.block_hash(subshare)
|
||||
#from allmydata.util import idlib
|
||||
#from allmydata.util import base32
|
||||
#log.msg("creating block (shareid=%d, blocknum=%d) "
|
||||
# "len=%d %r .. %r: %s" %
|
||||
# (shareid, segnum, len(subshare),
|
||||
# subshare[:50], subshare[-50:], idlib.b2a(subshare_hash)))
|
||||
# subshare[:50], subshare[-50:], base32.b2a(subshare_hash)))
|
||||
self.subshare_hashes[shareid].append(subshare_hash)
|
||||
|
||||
dl = self._gather_responses(dl)
|
||||
@ -518,7 +518,7 @@ class Encoder(object):
|
||||
d.addCallback(_got)
|
||||
def _got_hashtree_leaves(leaves):
|
||||
self.log("Encoder: got plaintext_hashtree_leaves: %s" %
|
||||
(",".join([idlib.b2a(h) for h in leaves]),),
|
||||
(",".join([base32.b2a(h) for h in leaves]),),
|
||||
level=log.NOISY)
|
||||
ht = list(HashTree(list(leaves)))
|
||||
self.uri_extension_data["plaintext_root_hash"] = ht[0]
|
||||
@ -636,7 +636,7 @@ class Encoder(object):
|
||||
ed = {}
|
||||
for k,v in self.uri_extension_data.items():
|
||||
if k.endswith("hash"):
|
||||
ed[k] = idlib.b2a(v)
|
||||
ed[k] = base32.b2a(v)
|
||||
else:
|
||||
ed[k] = v
|
||||
self.log("uri_extension_data is %s" % (ed,), level=log.NOISY, parent=lp)
|
||||
|
@ -48,7 +48,7 @@ or implied. It probably won't make your computer catch on fire,
|
||||
or eat your children, but it might. Use at your own risk.
|
||||
"""
|
||||
|
||||
from allmydata.util import idlib
|
||||
from allmydata.util import base32
|
||||
from allmydata.util.hashutil import tagged_hash, tagged_pair_hash
|
||||
|
||||
__version__ = '1.0.0-allmydata'
|
||||
@ -153,7 +153,7 @@ class CompleteBinaryTreeMixin:
|
||||
lines = []
|
||||
for i,depth in self.depth_first():
|
||||
lines.append("%s%3d: %s" % (" "*depth, i,
|
||||
idlib.b2a_or_none(self[i])))
|
||||
base32.b2a_or_none(self[i])))
|
||||
return "\n".join(lines) + "\n"
|
||||
|
||||
def get_leaf_index(self, leafnum):
|
||||
|
@ -6,7 +6,7 @@ from twisted.internet import defer
|
||||
from twisted.python import failure
|
||||
from foolscap.eventual import eventually
|
||||
from allmydata.interfaces import IMutableFileNode, IMutableFileURI
|
||||
from allmydata.util import hashutil, mathutil, idlib, log
|
||||
from allmydata.util import base32, hashutil, mathutil, idlib, log
|
||||
from allmydata.uri import WriteableSSKFileURI
|
||||
from allmydata import hashtree, codec, storage
|
||||
from allmydata.encode import NotEnoughPeersError
|
||||
@ -404,7 +404,7 @@ class Retrieve:
|
||||
# ok, it's a valid verinfo. Add it to the list of validated
|
||||
# versions.
|
||||
self.log(" found valid version %d-%s from %s-sh%d: %d-%d/%d/%d"
|
||||
% (seqnum, idlib.b2a(root_hash)[:4],
|
||||
% (seqnum, base32.b2a(root_hash)[:4],
|
||||
idlib.shortnodeid_b2a(peerid), shnum,
|
||||
k, N, segsize, datalength))
|
||||
self._valid_versions[verinfo] = (prefix, DictOfSets())
|
||||
@ -562,7 +562,7 @@ class Retrieve:
|
||||
shares_s.append("#%d" % shnum)
|
||||
shares_s = ",".join(shares_s)
|
||||
self.log("_attempt_decode: version %d-%s, shares: %s" %
|
||||
(seqnum, idlib.b2a(root_hash)[:4], shares_s))
|
||||
(seqnum, base32.b2a(root_hash)[:4], shares_s))
|
||||
|
||||
# first, validate each share that we haven't validated yet. We use
|
||||
# self._valid_shares to remember which ones we've already checked.
|
||||
@ -963,7 +963,7 @@ class Publish:
|
||||
for oldplace in current_share_peers.get(shnum, []):
|
||||
(peerid, seqnum, R) = oldplace
|
||||
logmsg2.append("%s:#%d:R=%s" % (idlib.shortnodeid_b2a(peerid),
|
||||
seqnum, idlib.b2a(R)[:4]))
|
||||
seqnum, base32.b2a(R)[:4]))
|
||||
logmsg.append("sh%d on (%s)" % (shnum, "/".join(logmsg2)))
|
||||
self.log("sharemap: %s" % (", ".join(logmsg)), level=log.NOISY)
|
||||
self.log("we are planning to push new seqnum=#%d" % self._new_seqnum,
|
||||
@ -1126,7 +1126,7 @@ class Publish:
|
||||
for i in needed_hashes ] )
|
||||
root_hash = share_hash_tree[0]
|
||||
assert len(root_hash) == 32
|
||||
self.log("my new root_hash is %s" % idlib.b2a(root_hash))
|
||||
self.log("my new root_hash is %s" % base32.b2a(root_hash))
|
||||
|
||||
prefix = pack_prefix(seqnum, root_hash, IV,
|
||||
required_shares, total_shares,
|
||||
@ -1257,8 +1257,8 @@ class Publish:
|
||||
" shnum=%d: I thought they had #%d:R=%s,"
|
||||
" but testv reported #%d:R=%s" %
|
||||
(shnum,
|
||||
seqnum, idlib.b2a(root_hash)[:4],
|
||||
old_seqnum, idlib.b2a(old_root_hash)[:4]),
|
||||
seqnum, base32.b2a(root_hash)[:4],
|
||||
old_seqnum, base32.b2a(old_root_hash)[:4]),
|
||||
parent=lp, level=log.WEIRD)
|
||||
surprised = True
|
||||
if surprised:
|
||||
@ -1268,7 +1268,7 @@ class Publish:
|
||||
for shnum, places in dispatch_map.items():
|
||||
sent_to = [(idlib.shortnodeid_b2a(peerid),
|
||||
seqnum,
|
||||
idlib.b2a(root_hash)[:4])
|
||||
base32.b2a(root_hash)[:4])
|
||||
for (peerid,seqnum,root_hash) in places]
|
||||
self.log(" share %d sent to: %s" % (shnum, sent_to),
|
||||
level=log.NOISY)
|
||||
|
@ -98,7 +98,7 @@ def format_expiration_time(expiration_time):
|
||||
|
||||
def dump_mutable_share(config, out, err):
|
||||
from allmydata import storage
|
||||
from allmydata.util import idlib
|
||||
from allmydata.util import base32, idlib
|
||||
m = storage.MutableShareFile(config['filename'])
|
||||
f = open(config['filename'], "rb")
|
||||
WE, nodeid = m._read_write_enabler_and_nodeid(f)
|
||||
@ -118,7 +118,7 @@ def dump_mutable_share(config, out, err):
|
||||
print >>out
|
||||
print >>out, "Mutable slot found:"
|
||||
print >>out, " share_type: %s" % share_type
|
||||
print >>out, " write_enabler: %s" % idlib.b2a(WE)
|
||||
print >>out, " write_enabler: %s" % base32.b2a(WE)
|
||||
print >>out, " WE for nodeid: %s" % idlib.nodeid_b2a(nodeid)
|
||||
print >>out, " num_extra_leases: %d" % num_extra_leases
|
||||
print >>out, " container_size: %d" % container_size
|
||||
@ -130,8 +130,8 @@ def dump_mutable_share(config, out, err):
|
||||
print >>out, " ownerid: %d" % oid
|
||||
when = format_expiration_time(et)
|
||||
print >>out, " expires in %s" % when
|
||||
print >>out, " renew_secret: %s" % idlib.b2a(rs)
|
||||
print >>out, " cancel_secret: %s" % idlib.b2a(cs)
|
||||
print >>out, " renew_secret: %s" % base32.b2a(rs)
|
||||
print >>out, " cancel_secret: %s" % base32.b2a(cs)
|
||||
print >>out, " secrets are for nodeid: %s" % idlib.nodeid_b2a(anid)
|
||||
else:
|
||||
print >>out, "No leases."
|
||||
@ -144,7 +144,7 @@ def dump_mutable_share(config, out, err):
|
||||
|
||||
def dump_SDMF_share(offset, length, config, out, err):
|
||||
from allmydata import mutable
|
||||
from allmydata.util import idlib
|
||||
from allmydata.util import base32
|
||||
|
||||
f = open(config['filename'], "rb")
|
||||
f.seek(offset)
|
||||
@ -168,8 +168,8 @@ def dump_SDMF_share(offset, length, config, out, err):
|
||||
|
||||
print >>out, " SDMF contents:"
|
||||
print >>out, " seqnum: %d" % seqnum
|
||||
print >>out, " root_hash: %s" % idlib.b2a(root_hash)
|
||||
print >>out, " IV: %s" % idlib.b2a(IV)
|
||||
print >>out, " root_hash: %s" % base32.b2a(root_hash)
|
||||
print >>out, " IV: %s" % base32.b2a(IV)
|
||||
print >>out, " required_shares: %d" % k
|
||||
print >>out, " total_shares: %d" % N
|
||||
print >>out, " segsize: %d" % segsize
|
||||
@ -194,7 +194,7 @@ class DumpCapOptions(usage.Options):
|
||||
|
||||
def dump_cap(config, out=sys.stdout, err=sys.stderr):
|
||||
from allmydata import uri
|
||||
from allmydata.util.idlib import a2b
|
||||
from allmydata.util import base32
|
||||
from base64 import b32decode
|
||||
import urlparse, urllib
|
||||
|
||||
@ -204,11 +204,11 @@ def dump_cap(config, out=sys.stdout, err=sys.stderr):
|
||||
nodeid = b32decode(config['nodeid'].upper())
|
||||
secret = None
|
||||
if config['client-secret']:
|
||||
secret = a2b(config['client-secret'])
|
||||
secret = base32.a2b(config['client-secret'])
|
||||
elif config['client-dir']:
|
||||
secretfile = os.path.join(config['client-dir'], "private", "secret")
|
||||
try:
|
||||
secret = a2b(open(secretfile, "r").read().strip())
|
||||
secret = base32.a2b(open(secretfile, "r").read().strip())
|
||||
except EnvironmentError:
|
||||
pass
|
||||
|
||||
@ -224,34 +224,33 @@ def dump_cap(config, out=sys.stdout, err=sys.stderr):
|
||||
|
||||
def _dump_secrets(storage_index, secret, nodeid, out):
|
||||
from allmydata.util import hashutil
|
||||
from allmydata.util.idlib import b2a
|
||||
from allmydata.util import base32
|
||||
|
||||
if secret:
|
||||
crs = hashutil.my_renewal_secret_hash(secret)
|
||||
print >>out, " client renewal secret:", b2a(crs)
|
||||
print >>out, " client renewal secret:", base32.b2a(crs)
|
||||
frs = hashutil.file_renewal_secret_hash(crs, storage_index)
|
||||
print >>out, " file renewal secret:", b2a(frs)
|
||||
print >>out, " file renewal secret:", base32.b2a(frs)
|
||||
if nodeid:
|
||||
renew = hashutil.bucket_renewal_secret_hash(frs, nodeid)
|
||||
print >>out, " lease renewal secret:", b2a(renew)
|
||||
print >>out, " lease renewal secret:", base32.b2a(renew)
|
||||
ccs = hashutil.my_cancel_secret_hash(secret)
|
||||
print >>out, " client cancel secret:", b2a(ccs)
|
||||
print >>out, " client cancel secret:", base32.b2a(ccs)
|
||||
fcs = hashutil.file_cancel_secret_hash(ccs, storage_index)
|
||||
print >>out, " file cancel secret:", b2a(fcs)
|
||||
print >>out, " file cancel secret:", base32.b2a(fcs)
|
||||
if nodeid:
|
||||
cancel = hashutil.bucket_cancel_secret_hash(fcs, nodeid)
|
||||
print >>out, " lease cancel secret:", b2a(cancel)
|
||||
print >>out, " lease cancel secret:", base32.b2a(cancel)
|
||||
|
||||
def dump_uri_instance(u, nodeid, secret, out, err, show_header=True):
|
||||
from allmydata import storage, uri
|
||||
from allmydata.util.idlib import b2a
|
||||
from allmydata.util import hashutil
|
||||
from allmydata.util import base32, hashutil
|
||||
|
||||
if isinstance(u, uri.CHKFileURI):
|
||||
if show_header:
|
||||
print >>out, "CHK File:"
|
||||
print >>out, " key:", b2a(u.key)
|
||||
print >>out, " UEB hash:", b2a(u.uri_extension_hash)
|
||||
print >>out, " key:", base32.b2a(u.key)
|
||||
print >>out, " UEB hash:", base32.b2a(u.uri_extension_hash)
|
||||
print >>out, " size:", u.size
|
||||
print >>out, " k/N: %d/%d" % (u.needed_shares, u.total_shares)
|
||||
print >>out, " storage index:", storage.si_b2a(u.storage_index)
|
||||
@ -259,7 +258,7 @@ def dump_uri_instance(u, nodeid, secret, out, err, show_header=True):
|
||||
elif isinstance(u, uri.CHKFileVerifierURI):
|
||||
if show_header:
|
||||
print >>out, "CHK Verifier URI:"
|
||||
print >>out, " UEB hash:", b2a(u.uri_extension_hash)
|
||||
print >>out, " UEB hash:", base32.b2a(u.uri_extension_hash)
|
||||
print >>out, " size:", u.size
|
||||
print >>out, " k/N: %d/%d" % (u.needed_shares, u.total_shares)
|
||||
print >>out, " storage index:", storage.si_b2a(u.storage_index)
|
||||
@ -272,28 +271,28 @@ def dump_uri_instance(u, nodeid, secret, out, err, show_header=True):
|
||||
elif isinstance(u, uri.WriteableSSKFileURI):
|
||||
if show_header:
|
||||
print >>out, "SSK Writeable URI:"
|
||||
print >>out, " writekey:", b2a(u.writekey)
|
||||
print >>out, " readkey:", b2a(u.readkey)
|
||||
print >>out, " writekey:", base32.b2a(u.writekey)
|
||||
print >>out, " readkey:", base32.b2a(u.readkey)
|
||||
print >>out, " storage index:", storage.si_b2a(u.storage_index)
|
||||
print >>out, " fingerprint:", b2a(u.fingerprint)
|
||||
print >>out, " fingerprint:", base32.b2a(u.fingerprint)
|
||||
print >>out
|
||||
if nodeid:
|
||||
we = hashutil.ssk_write_enabler_hash(u.writekey, nodeid)
|
||||
print >>out, " write_enabler:", b2a(we)
|
||||
print >>out, " write_enabler:", base32.b2a(we)
|
||||
print >>out
|
||||
_dump_secrets(u.storage_index, secret, nodeid, out)
|
||||
|
||||
elif isinstance(u, uri.ReadonlySSKFileURI):
|
||||
if show_header:
|
||||
print >>out, "SSK Read-only URI:"
|
||||
print >>out, " readkey:", b2a(u.readkey)
|
||||
print >>out, " readkey:", base32.b2a(u.readkey)
|
||||
print >>out, " storage index:", storage.si_b2a(u.storage_index)
|
||||
print >>out, " fingerprint:", b2a(u.fingerprint)
|
||||
print >>out, " fingerprint:", base32.b2a(u.fingerprint)
|
||||
elif isinstance(u, uri.SSKVerifierURI):
|
||||
if show_header:
|
||||
print >>out, "SSK Verifier URI:"
|
||||
print >>out, " storage index:", storage.si_b2a(u.storage_index)
|
||||
print >>out, " fingerprint:", b2a(u.fingerprint)
|
||||
print >>out, " fingerprint:", base32.b2a(u.fingerprint)
|
||||
|
||||
elif isinstance(u, uri.NewDirectoryURI):
|
||||
if show_header:
|
||||
@ -358,7 +357,7 @@ class CatalogSharesOptions(usage.Options):
|
||||
|
||||
def describe_share(abs_sharefile, si_s, shnum_s, now, out, err):
|
||||
from allmydata import uri, storage, mutable
|
||||
from allmydata.util import idlib
|
||||
from allmydata.util import base32
|
||||
import struct
|
||||
|
||||
f = open(abs_sharefile, "rb")
|
||||
@ -403,7 +402,7 @@ def describe_share(abs_sharefile, si_s, shnum_s, now, out, err):
|
||||
|
||||
print >>out, "SDMF %s %d/%d %d #%d:%s %d %s" % \
|
||||
(si_s, k, N, datalen,
|
||||
seqnum, idlib.b2a(root_hash),
|
||||
seqnum, base32.b2a(root_hash),
|
||||
expiration, abs_sharefile)
|
||||
else:
|
||||
print >>out, "UNKNOWN mutable %s" % (abs_sharefile,)
|
||||
|
@ -9,7 +9,7 @@ from zope.interface import implements
|
||||
from allmydata.interfaces import RIStorageServer, RIBucketWriter, \
|
||||
RIBucketReader, IStorageBucketWriter, IStorageBucketReader, HASH_SIZE, \
|
||||
BadWriteEnablerError, IStatsProducer
|
||||
from allmydata.util import base62, fileutil, idlib, mathutil, log
|
||||
from allmydata.util import base32, fileutil, idlib, mathutil, log
|
||||
from allmydata.util.assertutil import precondition, _assert
|
||||
import allmydata # for __version__
|
||||
|
||||
@ -48,10 +48,10 @@ NUM_RE=re.compile("^[0-9]+$")
|
||||
# B+0x48: next lease, or end of record
|
||||
|
||||
def si_b2a(storageindex):
|
||||
return base62.b2a(storageindex)
|
||||
return base32.b2a(storageindex)
|
||||
|
||||
def si_a2b(ascii_storageindex):
|
||||
return base62.a2b(ascii_storageindex)
|
||||
return base32.a2b(ascii_storageindex)
|
||||
|
||||
def storage_index_to_dir(storageindex):
|
||||
sia = si_b2a(storageindex)
|
||||
|
@ -89,31 +89,31 @@ class CLI(unittest.TestCase):
|
||||
size=size)
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("CHK File:" in output)
|
||||
self.failUnless("key: yyyoryarywdyqnyjbefoadeqbh" in output)
|
||||
self.failUnless("UEB hash: hd7rwri6djiapo6itg5hcxa7ze5im7z9qwcdu8oka6qinahsbiuo" in output)
|
||||
self.failUnless("key: aaaqeayeaudaocajbifqydiob4" in output, output)
|
||||
self.failUnless("UEB hash: 4d5euev6djvynq6vrg34mpy5xi3vl5x7oumdthqky6ovcy4wbvtq" in output, output)
|
||||
self.failUnless("size: 1234" in output)
|
||||
self.failUnless("k/N: 25/100" in output)
|
||||
self.failUnless("storage index: 2WlXTYP4ahK2VBkx1pckfC" in output, output)
|
||||
self.failUnless("storage index: kmkbjguwmkxej3wejdcvu74zki" in output, output)
|
||||
|
||||
output = self._dump_cap("--client-secret", "p3w849k9whqhw6b9fkf4xjs5xc",
|
||||
output = self._dump_cap("--client-secret", "5s33nk3qpvnj2fw3z4mnm2y6fa",
|
||||
u.to_string())
|
||||
self.failUnless("client renewal secret: pu3oy5fu4irjsudwhn6c71g87anrxi1kokt4hmxz7qh5p1895zpy" in output)
|
||||
self.failUnless("client renewal secret: jltcy6cppghq6ha3uzcawqr2lvwpzmw4teeqj2if6jd2vfpit6hq" in output, output)
|
||||
|
||||
output = self._dump_cap(u.get_verifier().to_string())
|
||||
self.failIf("key: " in output)
|
||||
self.failUnless("UEB hash: hd7rwri6djiapo6itg5hcxa7ze5im7z9qwcdu8oka6qinahsbiuo" in output)
|
||||
self.failUnless("UEB hash: 4d5euev6djvynq6vrg34mpy5xi3vl5x7oumdthqky6ovcy4wbvtq" in output, output)
|
||||
self.failUnless("size: 1234" in output)
|
||||
self.failUnless("k/N: 25/100" in output)
|
||||
self.failUnless("storage index: 2WlXTYP4ahK2VBkx1pckfC" in output, output)
|
||||
self.failUnless("storage index: kmkbjguwmkxej3wejdcvu74zki" in output, output)
|
||||
|
||||
prefixed_u = "http://127.0.0.1/uri/%s" % urllib.quote(u.to_string())
|
||||
output = self._dump_cap(prefixed_u)
|
||||
self.failUnless("CHK File:" in output)
|
||||
self.failUnless("key: yyyoryarywdyqnyjbefoadeqbh" in output)
|
||||
self.failUnless("UEB hash: hd7rwri6djiapo6itg5hcxa7ze5im7z9qwcdu8oka6qinahsbiuo" in output)
|
||||
self.failUnless("key: aaaqeayeaudaocajbifqydiob4" in output, output)
|
||||
self.failUnless("UEB hash: 4d5euev6djvynq6vrg34mpy5xi3vl5x7oumdthqky6ovcy4wbvtq" in output, output)
|
||||
self.failUnless("size: 1234" in output)
|
||||
self.failUnless("k/N: 25/100" in output)
|
||||
self.failUnless("storage index: 2WlXTYP4ahK2VBkx1pckfC" in output, output)
|
||||
self.failUnless("storage index: kmkbjguwmkxej3wejdcvu74zki" in output, output)
|
||||
|
||||
def test_dump_cap_lit(self):
|
||||
u = uri.LiteralFileURI("this is some data")
|
||||
@ -128,22 +128,22 @@ class CLI(unittest.TestCase):
|
||||
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("SSK Writeable URI:" in output)
|
||||
self.failUnless("writekey: yryonyebyryonyebyryonyebyr" in output)
|
||||
self.failUnless("readkey: zhgqsyrkuywo3rha41b1d7xrar" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("writekey: aeaqcaibaeaqcaibaeaqcaibae" in output, output)
|
||||
self.failUnless("readkey: x4gowaektauqze4y2sbsd5peye" in output, output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output, output)
|
||||
|
||||
output = self._dump_cap("--client-secret", "p3w849k9whqhw6b9fkf4xjs5xc",
|
||||
output = self._dump_cap("--client-secret", "tylkpgr364eav3ipsnq57yyafu",
|
||||
u.to_string())
|
||||
self.failUnless("file renewal secret: xy9p89q9pkitqn4ycwu5tpt9yia7s9izsqudnb4q5jdc3rawgcny" in output)
|
||||
self.failUnless("file renewal secret: cs54qwurfjmeduruapo46kqwexpcvav5oemczblonglj6xmoyvkq" in output, output)
|
||||
|
||||
fileutil.make_dirs("cli/test_dump_cap/private")
|
||||
f = open("cli/test_dump_cap/private/secret", "w")
|
||||
f.write("p3w849k9whqhw6b9fkf4xjs5xc\n")
|
||||
f.write("y6c7q34mjbt5kkf6hb3utuoj7u\n")
|
||||
f.close()
|
||||
output = self._dump_cap("--client-dir", "cli/test_dump_cap",
|
||||
u.to_string())
|
||||
self.failUnless("file renewal secret: xy9p89q9pkitqn4ycwu5tpt9yia7s9izsqudnb4q5jdc3rawgcny" in output)
|
||||
self.failUnless("file renewal secret: 4jkip4ie2zgmbhcni6g4vmsivwuakpbw7hwnmdancsc6fkrv27kq" in output, output)
|
||||
|
||||
output = self._dump_cap("--client-dir", "cli/test_dump_cap_BOGUS",
|
||||
u.to_string())
|
||||
@ -151,28 +151,28 @@ class CLI(unittest.TestCase):
|
||||
|
||||
output = self._dump_cap("--nodeid", "tqc35esocrvejvg4mablt6aowg6tl43j",
|
||||
u.to_string())
|
||||
self.failUnless("write_enabler: rqk9q6w46dim5ybshqk9kotkyhqcdqmp1z6498xniuz5kkjs1w7o" in output)
|
||||
self.failUnless("write_enabler: eok7o6u26dvl3abw4ok7kqrka4omdolnsx627hpcvtx3kkjwsu5q" in output, output)
|
||||
self.failIf("file renewal secret:" in output)
|
||||
|
||||
output = self._dump_cap("--nodeid", "tqc35esocrvejvg4mablt6aowg6tl43j",
|
||||
"--client-secret", "p3w849k9whqhw6b9fkf4xjs5xc",
|
||||
"--client-secret", "6orzlv22ggdhphjpmsixcbwufq",
|
||||
u.to_string())
|
||||
self.failUnless("write_enabler: rqk9q6w46dim5ybshqk9kotkyhqcdqmp1z6498xniuz5kkjs1w7o" in output)
|
||||
self.failUnless("file renewal secret: xy9p89q9pkitqn4ycwu5tpt9yia7s9izsqudnb4q5jdc3rawgcny" in output)
|
||||
self.failUnless("lease renewal secret: r3fsw67mfji3c9mtsisqdumc1pz3gquzdrh4cpu63h8du4uuedgo" in output)
|
||||
self.failUnless("write_enabler: eok7o6u26dvl3abw4ok7kqrka4omdolnsx627hpcvtx3kkjwsu5q" in output, output)
|
||||
self.failUnless("file renewal secret: aabhsp6kfsxb57jzdan4dnyzcd3m2prx34jd4z5nj5t5a7guf5fq" in output, output)
|
||||
self.failUnless("lease renewal secret: bajcslergse474ga775msalmxxapgwr27lngeja4u7ef5j7yh4bq" in output, output)
|
||||
|
||||
u = u.get_readonly()
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("SSK Read-only URI:" in output)
|
||||
self.failUnless("readkey: zhgqsyrkuywo3rha41b1d7xrar" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("readkey: x4gowaektauqze4y2sbsd5peye" in output, output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output)
|
||||
|
||||
u = u.get_verifier()
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("SSK Verifier URI:" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output)
|
||||
|
||||
def test_dump_cap_directory(self):
|
||||
writekey = "\x01" * 16
|
||||
@ -182,37 +182,37 @@ class CLI(unittest.TestCase):
|
||||
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("Directory Writeable URI:" in output)
|
||||
self.failUnless("writekey: yryonyebyryonyebyryonyebyr" in output)
|
||||
self.failUnless("readkey: zhgqsyrkuywo3rha41b1d7xrar" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("writekey: aeaqcaibaeaqcaibaeaqcaibae" in output, output)
|
||||
self.failUnless("readkey: x4gowaektauqze4y2sbsd5peye" in output, output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output, output)
|
||||
|
||||
output = self._dump_cap("--client-secret", "p3w849k9whqhw6b9fkf4xjs5xc",
|
||||
output = self._dump_cap("--client-secret", "a3nyfbnkorp377jhguslgc2dqi",
|
||||
u.to_string())
|
||||
self.failUnless("file renewal secret: xy9p89q9pkitqn4ycwu5tpt9yia7s9izsqudnb4q5jdc3rawgcny" in output)
|
||||
self.failUnless("file renewal secret: zwmq2azrd7lfcmhkrhpgjsxeb2vfpixgvrczbo2asqzdfbmiemwq" in output, output)
|
||||
|
||||
output = self._dump_cap("--nodeid", "tqc35esocrvejvg4mablt6aowg6tl43j",
|
||||
u.to_string())
|
||||
self.failUnless("write_enabler: rqk9q6w46dim5ybshqk9kotkyhqcdqmp1z6498xniuz5kkjs1w7o" in output)
|
||||
self.failUnless("write_enabler: eok7o6u26dvl3abw4ok7kqrka4omdolnsx627hpcvtx3kkjwsu5q" in output, output)
|
||||
self.failIf("file renewal secret:" in output)
|
||||
|
||||
output = self._dump_cap("--nodeid", "tqc35esocrvejvg4mablt6aowg6tl43j",
|
||||
"--client-secret", "p3w849k9whqhw6b9fkf4xjs5xc",
|
||||
"--client-secret", "rzaq5to2xm6e5otctpdvzw6bfa",
|
||||
u.to_string())
|
||||
self.failUnless("write_enabler: rqk9q6w46dim5ybshqk9kotkyhqcdqmp1z6498xniuz5kkjs1w7o" in output)
|
||||
self.failUnless("file renewal secret: xy9p89q9pkitqn4ycwu5tpt9yia7s9izsqudnb4q5jdc3rawgcny" in output)
|
||||
self.failUnless("lease renewal secret: r3fsw67mfji3c9mtsisqdumc1pz3gquzdrh4cpu63h8du4uuedgo" in output)
|
||||
self.failUnless("write_enabler: eok7o6u26dvl3abw4ok7kqrka4omdolnsx627hpcvtx3kkjwsu5q" in output, output)
|
||||
self.failUnless("file renewal secret: wdmu6rwefvmp2venbb4xz5u3273oybmuu553mi7uic37gfu6bacq" in output, output)
|
||||
self.failUnless("lease renewal secret: tlvwfudyfeqyss5kybt6ya72foedqxdovumlbt6ok7u5pyrf2mfq" in output, output)
|
||||
|
||||
u = u.get_readonly()
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("Directory Read-only URI:" in output)
|
||||
self.failUnless("readkey: zhgqsyrkuywo3rha41b1d7xrar" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("readkey: x4gowaektauqze4y2sbsd5peye" in output, output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output)
|
||||
|
||||
u = u.get_verifier()
|
||||
output = self._dump_cap(u.to_string())
|
||||
self.failUnless("Directory Verifier URI:" in output)
|
||||
self.failUnless("storage index: 4GWqxTUinIqKqWj770lRIA" in output, output)
|
||||
self.failUnless("fingerprint: 959x79z6959x79z6959x79z6959x79z6959x79z6959x79z6959y" in output)
|
||||
self.failUnless("storage index: rqx7xnpexjxuqprto6pezagdxi" in output, output)
|
||||
self.failUnless("fingerprint: 737p57x6737p57x6737p57x6737p57x6737p57x6737p57x6737a" in output, output)
|
||||
|
||||
|
@ -7,7 +7,7 @@ from twisted.python import log
|
||||
|
||||
import allmydata
|
||||
from allmydata import client, introducer
|
||||
from allmydata.util import idlib
|
||||
from allmydata.util import base32
|
||||
from foolscap.eventual import flushEventualQueue
|
||||
|
||||
class FakeIntroducerClient(introducer.IntroducerClient):
|
||||
@ -42,9 +42,9 @@ class Basic(unittest.TestCase):
|
||||
secret_fname = os.path.join(basedir, "private", "secret")
|
||||
self.failUnless(os.path.exists(secret_fname), secret_fname)
|
||||
renew_secret = c.get_renewal_secret()
|
||||
self.failUnless(idlib.b2a(renew_secret))
|
||||
self.failUnless(base32.b2a(renew_secret))
|
||||
cancel_secret = c.get_cancel_secret()
|
||||
self.failUnless(idlib.b2a(cancel_secret))
|
||||
self.failUnless(base32.b2a(cancel_secret))
|
||||
|
||||
def test_sizelimit_1(self):
|
||||
basedir = "client.Basic.test_sizelimit_1"
|
||||
|
@ -44,7 +44,7 @@ class Literal(unittest.TestCase):
|
||||
class Compare(unittest.TestCase):
|
||||
def test_compare(self):
|
||||
lit1 = uri.LiteralFileURI("some data")
|
||||
fileURI = 'URI:CHK:f3mf6az85wpcai8ma4qayfmxuc:nnw518w5hu3t5oohwtp7ah9n81z9rfg6c1ywk33ia3m64o67nsgo:3:10:345834'
|
||||
fileURI = 'URI:CHK:f5ahxa25t4qkktywz6teyfvcx4:opuioq7tj2y6idzfp6cazehtmgs5fdcebcz3cygrxyydvcozrmeq:3:10:345834'
|
||||
chk1 = uri.CHKFileURI.init_from_string(fileURI)
|
||||
chk2 = uri.CHKFileURI.init_from_string(fileURI)
|
||||
self.failIfEqual(lit1, chk1)
|
||||
@ -167,13 +167,13 @@ class Invalid(unittest.TestCase):
|
||||
|
||||
class Constraint(unittest.TestCase):
|
||||
def test_constraint(self):
|
||||
good="http://127.0.0.1:8123/uri/URI%3ADIR2%3Aqo8ayna47cpw3rx3kho3mu7q4h%3Abk9qbgx76gh6eyj5ps8p6buz8fffw1ofc37e9w9d6ncsfpuz7icy/"
|
||||
good="http://127.0.0.1:8123/uri/URI%3ADIR2%3Agh3l5rbvnv2333mrfvalmjfr4i%3Alz6l7u3z3b7g37s4zkdmfpx5ly4ib4m6thrpbusi6ys62qtc6mma/"
|
||||
uri.NewDirectoryURI.init_from_human_encoding(good)
|
||||
self.failUnlessRaises(AssertionError, uri.NewDirectoryURI.init_from_string, good)
|
||||
bad = good + '==='
|
||||
self.failUnlessRaises(AssertionError, uri.NewDirectoryURI.init_from_human_encoding, bad)
|
||||
self.failUnlessRaises(AssertionError, uri.NewDirectoryURI.init_from_string, bad)
|
||||
fileURI = 'URI:CHK:f3mf6az85wpcai8ma4qayfmxuc:nnw518w5hu3t5oohwtp7ah9n81z9rfg6c1ywk33ia3m64o67nsgo:3:10:345834'
|
||||
fileURI = 'URI:CHK:gh3l5rbvnv2333mrfvalmjfr4i:lz6l7u3z3b7g37s4zkdmfpx5ly4ib4m6thrpbusi6ys62qtc6mma:3:10:345834'
|
||||
uri.CHKFileURI.init_from_string(fileURI)
|
||||
|
||||
class Mutable(unittest.TestCase):
|
||||
|
@ -6,19 +6,21 @@ from twisted.trial import unittest
|
||||
from twisted.internet import defer
|
||||
from twisted.python import failure
|
||||
|
||||
from allmydata.util import bencode, idlib, humanreadable, mathutil, hashutil
|
||||
from allmydata.util import base32, bencode, idlib, humanreadable, mathutil, hashutil
|
||||
from allmydata.util import assertutil, fileutil, testutil, deferredutil
|
||||
|
||||
|
||||
class IDLib(unittest.TestCase):
|
||||
class Base32(unittest.TestCase):
|
||||
def test_b2a(self):
|
||||
self.failUnlessEqual(idlib.b2a("\x12\x34"), "ne4y")
|
||||
self.failUnlessEqual(base32.b2a("\x12\x34"), "ci2a")
|
||||
def test_b2a_or_none(self):
|
||||
self.failUnlessEqual(idlib.b2a_or_none(None), None)
|
||||
self.failUnlessEqual(idlib.b2a_or_none("\x12\x34"), "ne4y")
|
||||
self.failUnlessEqual(base32.b2a_or_none(None), None)
|
||||
self.failUnlessEqual(base32.b2a_or_none("\x12\x34"), "ci2a")
|
||||
def test_a2b(self):
|
||||
self.failUnlessEqual(idlib.a2b("ne4y"), "\x12\x34")
|
||||
self.failUnlessRaises(AssertionError, idlib.a2b, "b0gus")
|
||||
self.failUnlessEqual(base32.a2b("ci2a"), "\x12\x34")
|
||||
self.failUnlessRaises(AssertionError, base32.a2b, "b0gus")
|
||||
|
||||
class IDLib(unittest.TestCase):
|
||||
def test_nodeid_b2a(self):
|
||||
self.failUnlessEqual(idlib.nodeid_b2a("\x00"*20), "a"*32)
|
||||
|
||||
|
@ -13,7 +13,7 @@ from allmydata.util.hashutil import file_renewal_secret_hash, \
|
||||
bucket_cancel_secret_hash, plaintext_hasher, \
|
||||
storage_index_hash, plaintext_segment_hasher, content_hash_key_hasher
|
||||
from allmydata import encode, storage, hashtree, uri
|
||||
from allmydata.util import idlib, mathutil
|
||||
from allmydata.util import base32, idlib, mathutil
|
||||
from allmydata.util.assertutil import precondition
|
||||
from allmydata.interfaces import IUploadable, IUploader, IUploadResults, \
|
||||
IEncryptedUploadable, RIEncryptedUploadable, IUploadStatus
|
||||
@ -448,7 +448,7 @@ class EncryptAnUploadable:
|
||||
level=log.NOISY)
|
||||
self.log(format="plaintext leaf hash [%(segnum)d] is %(hash)s",
|
||||
segnum=len(self._plaintext_segment_hashes)-1,
|
||||
hash=idlib.b2a(p.digest()),
|
||||
hash=base32.b2a(p.digest()),
|
||||
level=log.NOISY)
|
||||
|
||||
offset += this_segment
|
||||
@ -547,7 +547,7 @@ class EncryptAnUploadable:
|
||||
level=log.NOISY)
|
||||
self.log(format="plaintext leaf hash [%(segnum)d] is %(hash)s",
|
||||
segnum=len(self._plaintext_segment_hashes)-1,
|
||||
hash=idlib.b2a(p.digest()),
|
||||
hash=base32.b2a(p.digest()),
|
||||
level=log.NOISY)
|
||||
assert len(self._plaintext_segment_hashes) == num_segments
|
||||
return defer.succeed(tuple(self._plaintext_segment_hashes[first:last]))
|
||||
|
@ -3,7 +3,7 @@ import re, urllib
|
||||
from zope.interface import implements
|
||||
from twisted.python.components import registerAdapter
|
||||
from allmydata import storage
|
||||
from allmydata.util import base62, idlib, hashutil
|
||||
from allmydata.util import base32, base32, hashutil
|
||||
from allmydata.interfaces import IURI, IDirnodeURI, IFileURI, IVerifierURI, \
|
||||
IMutableFileURI, INewDirectoryURI, IReadonlyNewDirectoryURI
|
||||
|
||||
@ -11,9 +11,8 @@ from allmydata.interfaces import IURI, IDirnodeURI, IFileURI, IVerifierURI, \
|
||||
# enough information to retrieve and validate the contents. It shall be
|
||||
# expressed in a limited character set (namely [TODO]).
|
||||
|
||||
ZBASE32STR_128bits = '(%s{25}%s)' % (idlib.ZBASE32CHAR, idlib.ZBASE32CHAR_3bits)
|
||||
ZBASE32STR_256bits = '(%s{51}%s)' % (idlib.ZBASE32CHAR, idlib.ZBASE32CHAR_1bits)
|
||||
ZBASE62STR_128bits = '(%s{22})' % (base62.ZBASE62CHAR)
|
||||
BASE32STR_128bits = '(%s{25}%s)' % (base32.BASE32CHAR, base32.BASE32CHAR_3bits)
|
||||
BASE32STR_256bits = '(%s{51}%s)' % (base32.BASE32CHAR, base32.BASE32CHAR_1bits)
|
||||
|
||||
SEP='(?::|%3A)'
|
||||
NUMBER='([0-9]+)'
|
||||
@ -38,11 +37,11 @@ class _BaseURI:
|
||||
class CHKFileURI(_BaseURI):
|
||||
implements(IURI, IFileURI)
|
||||
|
||||
STRING_RE=re.compile('^URI:CHK:'+ZBASE32STR_128bits+':'+
|
||||
ZBASE32STR_256bits+':'+NUMBER+':'+NUMBER+':'+NUMBER+
|
||||
STRING_RE=re.compile('^URI:CHK:'+BASE32STR_128bits+':'+
|
||||
BASE32STR_256bits+':'+NUMBER+':'+NUMBER+':'+NUMBER+
|
||||
'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'CHK'+SEP+
|
||||
ZBASE32STR_128bits+SEP+ZBASE32STR_256bits+SEP+NUMBER+
|
||||
BASE32STR_128bits+SEP+BASE32STR_256bits+SEP+NUMBER+
|
||||
SEP+NUMBER+SEP+NUMBER+'$')
|
||||
|
||||
def __init__(self, key, uri_extension_hash, needed_shares, total_shares,
|
||||
@ -61,14 +60,14 @@ class CHKFileURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)),
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)),
|
||||
int(mo.group(3)), int(mo.group(4)), int(mo.group(5)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)),
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)),
|
||||
int(mo.group(3)), int(mo.group(4)), int(mo.group(5)))
|
||||
|
||||
def to_string(self):
|
||||
@ -77,8 +76,8 @@ class CHKFileURI(_BaseURI):
|
||||
assert isinstance(self.size, (int,long))
|
||||
|
||||
return ('URI:CHK:%s:%s:%d:%d:%d' %
|
||||
(idlib.b2a(self.key),
|
||||
idlib.b2a(self.uri_extension_hash),
|
||||
(base32.b2a(self.key),
|
||||
base32.b2a(self.uri_extension_hash),
|
||||
self.needed_shares,
|
||||
self.total_shares,
|
||||
self.size))
|
||||
@ -103,10 +102,10 @@ class CHKFileURI(_BaseURI):
|
||||
class CHKFileVerifierURI(_BaseURI):
|
||||
implements(IVerifierURI)
|
||||
|
||||
STRING_RE=re.compile('^URI:CHK-Verifier:'+ZBASE62STR_128bits+':'+
|
||||
ZBASE32STR_256bits+':'+NUMBER+':'+NUMBER+':'+NUMBER)
|
||||
STRING_RE=re.compile('^URI:CHK-Verifier:'+BASE32STR_128bits+':'+
|
||||
BASE32STR_256bits+':'+NUMBER+':'+NUMBER+':'+NUMBER)
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'CHK-Verifier'+SEP+
|
||||
ZBASE62STR_128bits+SEP+ZBASE32STR_256bits+SEP+NUMBER+
|
||||
BASE32STR_128bits+SEP+BASE32STR_256bits+SEP+NUMBER+
|
||||
SEP+NUMBER+SEP+NUMBER)
|
||||
|
||||
def __init__(self, storage_index, uri_extension_hash,
|
||||
@ -122,14 +121,14 @@ class CHKFileVerifierURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)),
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)),
|
||||
int(mo.group(3)), int(mo.group(4)), int(mo.group(5)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, (uri, cls, cls.STRING_RE)
|
||||
return cls(storage.si_a2b(mo.group(1)), idlib.a2b(mo.group(2)),
|
||||
return cls(storage.si_a2b(mo.group(1)), base32.a2b(mo.group(2)),
|
||||
int(mo.group(3)), int(mo.group(4)), int(mo.group(5)))
|
||||
|
||||
def to_string(self):
|
||||
@ -139,7 +138,7 @@ class CHKFileVerifierURI(_BaseURI):
|
||||
|
||||
return ('URI:CHK-Verifier:%s:%s:%d:%d:%d' %
|
||||
(storage.si_b2a(self.storage_index),
|
||||
idlib.b2a(self.uri_extension_hash),
|
||||
base32.b2a(self.uri_extension_hash),
|
||||
self.needed_shares,
|
||||
self.total_shares,
|
||||
self.size))
|
||||
@ -148,8 +147,8 @@ class CHKFileVerifierURI(_BaseURI):
|
||||
class LiteralFileURI(_BaseURI):
|
||||
implements(IURI, IFileURI)
|
||||
|
||||
STRING_RE=re.compile('^URI:LIT:'+idlib.ZBASE32STR_anybytes+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'LIT'+SEP+idlib.ZBASE32STR_anybytes+'$')
|
||||
STRING_RE=re.compile('^URI:LIT:'+base32.BASE32STR_anybytes+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'LIT'+SEP+base32.BASE32STR_anybytes+'$')
|
||||
|
||||
def __init__(self, data=None):
|
||||
if data is not None:
|
||||
@ -159,16 +158,16 @@ class LiteralFileURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)))
|
||||
return cls(base32.a2b(mo.group(1)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)))
|
||||
return cls(base32.a2b(mo.group(1)))
|
||||
|
||||
def to_string(self):
|
||||
return 'URI:LIT:%s' % idlib.b2a(self.data)
|
||||
return 'URI:LIT:%s' % base32.b2a(self.data)
|
||||
|
||||
def is_readonly(self):
|
||||
return True
|
||||
@ -188,10 +187,10 @@ class WriteableSSKFileURI(_BaseURI):
|
||||
implements(IURI, IMutableFileURI)
|
||||
|
||||
BASE_STRING='URI:SSK:'
|
||||
STRING_RE=re.compile('^'+BASE_STRING+ZBASE32STR_128bits+':'+
|
||||
ZBASE32STR_256bits+'$')
|
||||
STRING_RE=re.compile('^'+BASE_STRING+BASE32STR_128bits+':'+
|
||||
BASE32STR_256bits+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'SSK'+SEP+
|
||||
ZBASE32STR_128bits+SEP+ZBASE32STR_256bits+'$')
|
||||
BASE32STR_128bits+SEP+BASE32STR_256bits+'$')
|
||||
|
||||
def __init__(self, writekey, fingerprint):
|
||||
self.writekey = writekey
|
||||
@ -204,25 +203,25 @@ class WriteableSSKFileURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, (uri, cls)
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
def to_string(self):
|
||||
assert isinstance(self.writekey, str)
|
||||
assert isinstance(self.fingerprint, str)
|
||||
return 'URI:SSK:%s:%s' % (idlib.b2a(self.writekey),
|
||||
idlib.b2a(self.fingerprint))
|
||||
return 'URI:SSK:%s:%s' % (base32.b2a(self.writekey),
|
||||
base32.b2a(self.fingerprint))
|
||||
|
||||
def __repr__(self):
|
||||
return "<%s %s>" % (self.__class__.__name__, self.abbrev())
|
||||
|
||||
def abbrev(self):
|
||||
return idlib.b2a(self.writekey[:5])
|
||||
return base32.b2a(self.writekey[:5])
|
||||
|
||||
def is_readonly(self):
|
||||
return False
|
||||
@ -237,8 +236,8 @@ class ReadonlySSKFileURI(_BaseURI):
|
||||
implements(IURI, IMutableFileURI)
|
||||
|
||||
BASE_STRING='URI:SSK-RO:'
|
||||
STRING_RE=re.compile('^URI:SSK-RO:'+ZBASE32STR_128bits+':'+ZBASE32STR_256bits+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'SSK-RO'+SEP+ZBASE32STR_128bits+SEP+ZBASE32STR_256bits+'$')
|
||||
STRING_RE=re.compile('^URI:SSK-RO:'+BASE32STR_128bits+':'+BASE32STR_256bits+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'SSK-RO'+SEP+BASE32STR_128bits+SEP+BASE32STR_256bits+'$')
|
||||
|
||||
def __init__(self, readkey, fingerprint):
|
||||
self.readkey = readkey
|
||||
@ -250,25 +249,25 @@ class ReadonlySSKFileURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(idlib.a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(base32.a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
def to_string(self):
|
||||
assert isinstance(self.readkey, str)
|
||||
assert isinstance(self.fingerprint, str)
|
||||
return 'URI:SSK-RO:%s:%s' % (idlib.b2a(self.readkey),
|
||||
idlib.b2a(self.fingerprint))
|
||||
return 'URI:SSK-RO:%s:%s' % (base32.b2a(self.readkey),
|
||||
base32.b2a(self.fingerprint))
|
||||
|
||||
def __repr__(self):
|
||||
return "<%s %s>" % (self.__class__.__name__, self.abbrev())
|
||||
|
||||
def abbrev(self):
|
||||
return idlib.b2a(self.readkey[:5])
|
||||
return base32.b2a(self.readkey[:5])
|
||||
|
||||
def is_readonly(self):
|
||||
return True
|
||||
@ -283,8 +282,8 @@ class SSKVerifierURI(_BaseURI):
|
||||
implements(IVerifierURI)
|
||||
|
||||
BASE_STRING='URI:SSK-Verifier:'
|
||||
STRING_RE=re.compile('^'+BASE_STRING+ZBASE62STR_128bits+':'+ZBASE32STR_256bits+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'SSK-RO'+SEP+ZBASE62STR_128bits+SEP+ZBASE32STR_256bits+'$')
|
||||
STRING_RE=re.compile('^'+BASE_STRING+BASE32STR_128bits+':'+BASE32STR_256bits+'$')
|
||||
HUMAN_RE=re.compile('^'+OPTIONALHTTPLEAD+'URI'+SEP+'SSK-RO'+SEP+BASE32STR_128bits+SEP+BASE32STR_256bits+'$')
|
||||
|
||||
def __init__(self, storage_index, fingerprint):
|
||||
assert len(storage_index) == 16
|
||||
@ -295,19 +294,19 @@ class SSKVerifierURI(_BaseURI):
|
||||
def init_from_human_encoding(cls, uri):
|
||||
mo = cls.HUMAN_RE.search(uri)
|
||||
assert mo, uri
|
||||
return cls(storage.si_a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(storage.si_a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
@classmethod
|
||||
def init_from_string(cls, uri):
|
||||
mo = cls.STRING_RE.search(uri)
|
||||
assert mo, (uri, cls)
|
||||
return cls(storage.si_a2b(mo.group(1)), idlib.a2b(mo.group(2)))
|
||||
return cls(storage.si_a2b(mo.group(1)), base32.a2b(mo.group(2)))
|
||||
|
||||
def to_string(self):
|
||||
assert isinstance(self.storage_index, str)
|
||||
assert isinstance(self.fingerprint, str)
|
||||
return 'URI:SSK-Verifier:%s:%s' % (storage.si_b2a(self.storage_index),
|
||||
idlib.b2a(self.fingerprint))
|
||||
base32.b2a(self.fingerprint))
|
||||
|
||||
class _NewDirectoryBaseURI(_BaseURI):
|
||||
implements(IURI, IDirnodeURI)
|
||||
@ -513,6 +512,6 @@ def unpack_extension_readable(data):
|
||||
unpacked["UEB_hash"] = hashutil.uri_extension_hash(data)
|
||||
for k in sorted(unpacked.keys()):
|
||||
if 'hash' in k:
|
||||
unpacked[k] = idlib.b2a(unpacked[k])
|
||||
unpacked[k] = base32.b2a(unpacked[k])
|
||||
return unpacked
|
||||
|
||||
|
261
src/allmydata/util/base32.py
Normal file
261
src/allmydata/util/base32.py
Normal file
@ -0,0 +1,261 @@
|
||||
# from the Python Standard Library
|
||||
import string
|
||||
|
||||
from assertutil import precondition
|
||||
|
||||
z_base_32_alphabet = "ybndrfg8ejkmcpqxot1uwisza345h769" # Zooko's choice, rationale in "DESIGN" doc
|
||||
rfc3548_alphabet = "abcdefghijklmnopqrstuvwxyz234567" # RFC3548 standard used by Gnutella, Content-Addressable Web, THEX, Bitzi, Web-Calculus...
|
||||
chars = rfc3548_alphabet
|
||||
|
||||
vals = ''.join(map(chr, range(32)))
|
||||
c2vtranstable = string.maketrans(chars, vals)
|
||||
v2ctranstable = string.maketrans(vals, chars)
|
||||
identitytranstable = string.maketrans(chars, chars)
|
||||
|
||||
def _get_trailing_chars_without_lsbs(N, d):
|
||||
"""
|
||||
@return: a list of chars that can legitimately appear in the last place when the least significant N bits are ignored.
|
||||
"""
|
||||
s = []
|
||||
if N < 4:
|
||||
s.extend(_get_trailing_chars_without_lsbs(N+1, d=d))
|
||||
i = 0
|
||||
while i < len(chars):
|
||||
if not d.has_key(i):
|
||||
d[i] = None
|
||||
s.append(chars[i])
|
||||
i = i + 2**N
|
||||
return s
|
||||
|
||||
def get_trailing_chars_without_lsbs(N):
|
||||
precondition((N >= 0) and (N < 5), "N is required to be > 0 and < len(chars).", N=N)
|
||||
if N == 0:
|
||||
return chars
|
||||
d = {}
|
||||
return ''.join(_get_trailing_chars_without_lsbs(N, d=d))
|
||||
|
||||
BASE32CHAR = '['+get_trailing_chars_without_lsbs(0)+']'
|
||||
BASE32CHAR_4bits = '['+get_trailing_chars_without_lsbs(1)+']'
|
||||
BASE32CHAR_3bits = '['+get_trailing_chars_without_lsbs(2)+']'
|
||||
BASE32CHAR_2bits = '['+get_trailing_chars_without_lsbs(3)+']'
|
||||
BASE32CHAR_1bits = '['+get_trailing_chars_without_lsbs(4)+']'
|
||||
BASE32STR_1byte = BASE32CHAR+BASE32CHAR_3bits
|
||||
BASE32STR_2bytes = BASE32CHAR+'{3}'+BASE32CHAR_1bits
|
||||
BASE32STR_3bytes = BASE32CHAR+'{4}'+BASE32CHAR_4bits
|
||||
BASE32STR_4bytes = BASE32CHAR+'{6}'+BASE32CHAR_2bits
|
||||
BASE32STR_anybytes = '((?:%s{8})*' % (BASE32CHAR,) + "(?:|%s|%s|%s|%s))" % (BASE32STR_1byte, BASE32STR_2bytes, BASE32STR_3bytes, BASE32STR_4bytes)
|
||||
|
||||
def b2a(os):
|
||||
"""
|
||||
@param os the data to be encoded (a string)
|
||||
|
||||
@return the contents of os in base-32 encoded form
|
||||
"""
|
||||
return b2a_l(os, len(os)*8)
|
||||
|
||||
def b2a_or_none(os):
|
||||
if os is not None:
|
||||
return b2a(os)
|
||||
|
||||
def b2a_l(os, lengthinbits):
|
||||
"""
|
||||
@param os the data to be encoded (a string)
|
||||
@param lengthinbits the number of bits of data in os to be encoded
|
||||
|
||||
b2a_l() will generate a base-32 encoded string big enough to encode lengthinbits bits. So for
|
||||
example if os is 2 bytes long and lengthinbits is 15, then b2a_l() will generate a 3-character-
|
||||
long base-32 encoded string (since 3 quintets is sufficient to encode 15 bits). If os is
|
||||
2 bytes long and lengthinbits is 16 (or None), then b2a_l() will generate a 4-character string.
|
||||
Note that b2a_l() does not mask off unused least-significant bits, so for example if os is
|
||||
2 bytes long and lengthinbits is 15, then you must ensure that the unused least-significant bit
|
||||
of os is a zero bit or you will get the wrong result. This precondition is tested by assertions
|
||||
if assertions are enabled.
|
||||
|
||||
Warning: if you generate a base-32 encoded string with b2a_l(), and then someone else tries to
|
||||
decode it by calling a2b() instead of a2b_l(), then they will (probably) get a different
|
||||
string than the one you encoded! So only use b2a_l() when you are sure that the encoding and
|
||||
decoding sides know exactly which lengthinbits to use. If you do not have a way for the
|
||||
encoder and the decoder to agree upon the lengthinbits, then it is best to use b2a() and
|
||||
a2b(). The only drawback to using b2a() over b2a_l() is that when you have a number of
|
||||
bits to encode that is not a multiple of 8, b2a() can sometimes generate a base-32 encoded
|
||||
string that is one or two characters longer than necessary.
|
||||
|
||||
@return the contents of os in base-32 encoded form
|
||||
"""
|
||||
precondition(isinstance(lengthinbits, (int, long,)), "lengthinbits is required to be an integer.", lengthinbits=lengthinbits)
|
||||
precondition((lengthinbits+7)/8 == len(os), "lengthinbits is required to specify a number of bits storable in exactly len(os) octets.", lengthinbits=lengthinbits, lenos=len(os))
|
||||
|
||||
os = map(ord, os)
|
||||
|
||||
numquintets = (lengthinbits+4)/5
|
||||
numoctetsofdata = (lengthinbits+7)/8
|
||||
# print "numoctetsofdata: %s, len(os): %s, lengthinbits: %s, numquintets: %s" % (numoctetsofdata, len(os), lengthinbits, numquintets,)
|
||||
# strip trailing octets that won't be used
|
||||
del os[numoctetsofdata:]
|
||||
# zero out any unused bits in the final octet
|
||||
if lengthinbits % 8 != 0:
|
||||
os[-1] = os[-1] >> (8-(lengthinbits % 8))
|
||||
os[-1] = os[-1] << (8-(lengthinbits % 8))
|
||||
# append zero octets for padding if needed
|
||||
numoctetsneeded = (numquintets*5+7)/8 + 1
|
||||
os.extend([0]*(numoctetsneeded-len(os)))
|
||||
|
||||
quintets = []
|
||||
cutoff = 256
|
||||
num = os[0]
|
||||
i = 0
|
||||
while len(quintets) < numquintets:
|
||||
i = i + 1
|
||||
assert len(os) > i, "len(os): %s, i: %s, len(quintets): %s, numquintets: %s, lengthinbits: %s, numoctetsofdata: %s, numoctetsneeded: %s, os: %s" % (len(os), i, len(quintets), numquintets, lengthinbits, numoctetsofdata, numoctetsneeded, os,)
|
||||
num = num * 256
|
||||
num = num + os[i]
|
||||
if cutoff == 1:
|
||||
cutoff = 256
|
||||
continue
|
||||
cutoff = cutoff * 8
|
||||
quintet = num / cutoff
|
||||
quintets.append(quintet)
|
||||
num = num - (quintet * cutoff)
|
||||
|
||||
cutoff = cutoff / 32
|
||||
quintet = num / cutoff
|
||||
quintets.append(quintet)
|
||||
num = num - (quintet * cutoff)
|
||||
|
||||
if len(quintets) > numquintets:
|
||||
assert len(quintets) == (numquintets+1), "len(quintets): %s, numquintets: %s, quintets: %s" % (len(quintets), numquintets, quintets,)
|
||||
quintets = quintets[:numquintets]
|
||||
res = string.translate(string.join(map(chr, quintets), ''), v2ctranstable)
|
||||
assert could_be_base32_encoded_l(res, lengthinbits), "lengthinbits: %s, res: %s" % (lengthinbits, res,)
|
||||
return res
|
||||
|
||||
# b2a() uses the minimal number of quintets sufficient to encode the binary
|
||||
# input. It just so happens that the relation is like this (everything is
|
||||
# modulo 40 bits).
|
||||
# num_qs = NUM_OS_TO_NUM_QS[num_os]
|
||||
NUM_OS_TO_NUM_QS=(0, 2, 4, 5, 7,)
|
||||
|
||||
# num_os = NUM_QS_TO_NUM_OS[num_qs], but if not NUM_QS_LEGIT[num_qs] then
|
||||
# there is *no* number of octets which would have resulted in this number of
|
||||
# quintets, so either the encoded string has been mangled (truncated) or else
|
||||
# you were supposed to decode it with a2b_l() (which means you were supposed
|
||||
# to know the actual length of the encoded data).
|
||||
|
||||
NUM_QS_TO_NUM_OS=(0, 1, 1, 2, 2, 3, 3, 4)
|
||||
NUM_QS_LEGIT=(1, 0, 1, 0, 1, 1, 0, 1,)
|
||||
NUM_QS_TO_NUM_BITS=tuple(map(lambda x: x*8, NUM_QS_TO_NUM_OS))
|
||||
|
||||
# A fast way to determine whether a given string *could* be base-32 encoded data, assuming that the
|
||||
# original data had 8K bits for a positive integer K.
|
||||
# The boolean value of s8[len(s)%8][ord(s[-1])], where s is the possibly base-32 encoded string
|
||||
# tells whether the final character is reasonable.
|
||||
def add_check_array(cs, sfmap):
|
||||
checka=[0] * 256
|
||||
for c in cs:
|
||||
checka[ord(c)] = 1
|
||||
sfmap.append(tuple(checka))
|
||||
|
||||
def init_s8():
|
||||
s8 = []
|
||||
add_check_array(chars, s8)
|
||||
for lenmod8 in (1, 2, 3, 4, 5, 6, 7,):
|
||||
if NUM_QS_LEGIT[lenmod8]:
|
||||
add_check_array(get_trailing_chars_without_lsbs(4-(NUM_QS_TO_NUM_BITS[lenmod8]%5)), s8)
|
||||
else:
|
||||
add_check_array('', s8)
|
||||
return tuple(s8)
|
||||
s8 = init_s8()
|
||||
|
||||
# A somewhat fast way to determine whether a given string *could* be base-32 encoded data, given a
|
||||
# lengthinbits.
|
||||
# The boolean value of s5[lengthinbits%5][ord(s[-1])], where s is the possibly base-32 encoded
|
||||
# string tells whether the final character is reasonable.
|
||||
def init_s5():
|
||||
s5 = []
|
||||
add_check_array(get_trailing_chars_without_lsbs(0), s5)
|
||||
for lenmod5 in [1,2,3,4]:
|
||||
add_check_array(get_trailing_chars_without_lsbs(5-lenmod5), s5)
|
||||
return tuple(s5)
|
||||
s5 = init_s5()
|
||||
|
||||
def could_be_base32_encoded(s, s8=s8, tr=string.translate, identitytranstable=identitytranstable, chars=chars):
|
||||
if s == '':
|
||||
return True
|
||||
return s8[len(s)%8][ord(s[-1])] and not tr(s, identitytranstable, chars)
|
||||
|
||||
def could_be_base32_encoded_l(s, lengthinbits, s5=s5, tr=string.translate, identitytranstable=identitytranstable, chars=chars):
|
||||
if s == '':
|
||||
return True
|
||||
assert lengthinbits%5 < len(s5), lengthinbits
|
||||
assert ord(s[-1]) < s5[lengthinbits%5]
|
||||
return (((lengthinbits+4)/5) == len(s)) and s5[lengthinbits%5][ord(s[-1])] and not string.translate(s, identitytranstable, chars)
|
||||
|
||||
def num_octets_that_encode_to_this_many_quintets(numqs):
|
||||
# Here is a computation that conveniently expresses this:
|
||||
return (numqs*5+3)/8
|
||||
|
||||
def a2b(cs):
|
||||
"""
|
||||
@param cs the base-32 encoded data (a string)
|
||||
"""
|
||||
precondition(could_be_base32_encoded(cs), "cs is required to be possibly base32 encoded data.", cs=cs)
|
||||
|
||||
return a2b_l(cs, num_octets_that_encode_to_this_many_quintets(len(cs))*8)
|
||||
|
||||
def a2b_l(cs, lengthinbits):
|
||||
"""
|
||||
@param lengthinbits the number of bits of data in encoded into cs
|
||||
|
||||
a2b_l() will return a result big enough to hold lengthinbits bits. So for example if cs is
|
||||
4 characters long (encoding at least 15 and up to 20 bits) and lengthinbits is 16, then a2b_l()
|
||||
will return a string of length 2 (since 2 bytes is sufficient to store 16 bits). If cs is 4
|
||||
characters long and lengthinbits is 20, then a2b_l() will return a string of length 3 (since
|
||||
3 bytes is sufficient to store 20 bits). Note that b2a_l() does not mask off unused least-
|
||||
significant bits, so for example if cs is 4 characters long and lengthinbits is 17, then you
|
||||
must ensure that all three of the unused least-significant bits of cs are zero bits or you will
|
||||
get the wrong result. This precondition is tested by assertions if assertions are enabled.
|
||||
(Generally you just require the encoder to ensure this consistency property between the least
|
||||
significant zero bits and value of lengthinbits, and reject strings that have a length-in-bits
|
||||
which isn't a multiple of 8 and yet don't have trailing zero bits, as improperly encoded.)
|
||||
|
||||
Please see the warning in the docstring of b2a_l() regarding the use of b2a() versus b2a_l().
|
||||
|
||||
@return the data encoded in cs
|
||||
"""
|
||||
precondition(could_be_base32_encoded_l(cs, lengthinbits), "cs is required to be possibly base32 encoded data.", cs=cs, lengthinbits=lengthinbits)
|
||||
if cs == '':
|
||||
return ''
|
||||
|
||||
qs = map(ord, string.translate(cs, c2vtranstable))
|
||||
|
||||
numoctets = (lengthinbits+7)/8
|
||||
numquintetsofdata = (lengthinbits+4)/5
|
||||
# strip trailing quintets that won't be used
|
||||
del qs[numquintetsofdata:]
|
||||
# zero out any unused bits in the final quintet
|
||||
if lengthinbits % 5 != 0:
|
||||
qs[-1] = qs[-1] >> (5-(lengthinbits % 5))
|
||||
qs[-1] = qs[-1] << (5-(lengthinbits % 5))
|
||||
# append zero quintets for padding if needed
|
||||
numquintetsneeded = (numoctets*8+4)/5
|
||||
qs.extend([0]*(numquintetsneeded-len(qs)))
|
||||
|
||||
octets = []
|
||||
pos = 2048
|
||||
num = qs[0] * pos
|
||||
readybits = 5
|
||||
i = 1
|
||||
while len(octets) < numoctets:
|
||||
while pos > 256:
|
||||
pos = pos / 32
|
||||
num = num + (qs[i] * pos)
|
||||
i = i + 1
|
||||
octet = num / 256
|
||||
octets.append(octet)
|
||||
num = num - (octet * 256)
|
||||
num = num * 256
|
||||
pos = pos * 256
|
||||
assert len(octets) == numoctets, "len(octets): %s, numoctets: %s, octets: %s" % (len(octets), numoctets, octets,)
|
||||
res = ''.join(map(chr, octets))
|
||||
precondition(b2a_l(res, lengthinbits) == cs, "cs is required to be the canonical base-32 encoding of some data.", b2a(res), res=res, cs=cs)
|
||||
return res
|
@ -13,7 +13,7 @@ from allmydata.util.mathutil import log_ceil, log_floor
|
||||
|
||||
chars = "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz"
|
||||
|
||||
ZBASE62CHAR = '[' + chars + ']'
|
||||
BASE62CHAR = '[' + chars + ']'
|
||||
|
||||
vals = ''.join([chr(i) for i in range(62)])
|
||||
c2vtranstable = string.maketrans(chars, vals)
|
||||
|
@ -1,264 +1,4 @@
|
||||
# from the Python Standard Library
|
||||
import string
|
||||
|
||||
from assertutil import precondition
|
||||
|
||||
z_base_32_alphabet = "ybndrfg8ejkmcpqxot1uwisza345h769" # Zooko's choice, rationale in "DESIGN" doc
|
||||
rfc3548_alphabet = "abcdefghijklmnopqrstuvwxyz234567" # RFC3548 standard used by Gnutella, Content-Addressable Web, THEX, Bitzi, Web-Calculus...
|
||||
chars = z_base_32_alphabet
|
||||
|
||||
vals = ''.join(map(chr, range(32)))
|
||||
c2vtranstable = string.maketrans(chars, vals)
|
||||
v2ctranstable = string.maketrans(vals, chars)
|
||||
identitytranstable = string.maketrans(chars, chars)
|
||||
|
||||
def _get_trailing_chars_without_lsbs(N, d):
|
||||
"""
|
||||
@return: a list of chars that can legitimately appear in the last place when the least significant N bits are ignored.
|
||||
"""
|
||||
s = []
|
||||
if N < 4:
|
||||
s.extend(_get_trailing_chars_without_lsbs(N+1, d=d))
|
||||
i = 0
|
||||
while i < len(chars):
|
||||
if not d.has_key(i):
|
||||
d[i] = None
|
||||
s.append(chars[i])
|
||||
i = i + 2**N
|
||||
return s
|
||||
|
||||
def get_trailing_chars_without_lsbs(N):
|
||||
precondition((N >= 0) and (N < 5), "N is required to be > 0 and < len(chars).", N=N)
|
||||
if N == 0:
|
||||
return chars
|
||||
d = {}
|
||||
return ''.join(_get_trailing_chars_without_lsbs(N, d=d))
|
||||
|
||||
ZBASE32CHAR = '['+get_trailing_chars_without_lsbs(0)+']'
|
||||
ZBASE32CHAR_4bits = '['+get_trailing_chars_without_lsbs(1)+']'
|
||||
ZBASE32CHAR_3bits = '['+get_trailing_chars_without_lsbs(2)+']'
|
||||
ZBASE32CHAR_2bits = '['+get_trailing_chars_without_lsbs(3)+']'
|
||||
ZBASE32CHAR_1bits = '['+get_trailing_chars_without_lsbs(4)+']'
|
||||
ZBASE32STR_1byte = ZBASE32CHAR+ZBASE32CHAR_3bits
|
||||
ZBASE32STR_2bytes = ZBASE32CHAR+'{3}'+ZBASE32CHAR_1bits
|
||||
ZBASE32STR_3bytes = ZBASE32CHAR+'{4}'+ZBASE32CHAR_4bits
|
||||
ZBASE32STR_4bytes = ZBASE32CHAR+'{6}'+ZBASE32CHAR_2bits
|
||||
ZBASE32STR_anybytes = '((?:%s{8})*' % (ZBASE32CHAR,) + "(?:|%s|%s|%s|%s))" % (ZBASE32STR_1byte, ZBASE32STR_2bytes, ZBASE32STR_3bytes, ZBASE32STR_4bytes)
|
||||
|
||||
def b2a(os):
|
||||
"""
|
||||
@param os the data to be encoded (a string)
|
||||
|
||||
@return the contents of os in base-32 encoded form
|
||||
"""
|
||||
return b2a_l(os, len(os)*8)
|
||||
|
||||
def b2a_or_none(os):
|
||||
if os is not None:
|
||||
return b2a(os)
|
||||
|
||||
def b2a_l(os, lengthinbits):
|
||||
"""
|
||||
@param os the data to be encoded (a string)
|
||||
@param lengthinbits the number of bits of data in os to be encoded
|
||||
|
||||
b2a_l() will generate a base-32 encoded string big enough to encode lengthinbits bits. So for
|
||||
example if os is 2 bytes long and lengthinbits is 15, then b2a_l() will generate a 3-character-
|
||||
long base-32 encoded string (since 3 quintets is sufficient to encode 15 bits). If os is
|
||||
2 bytes long and lengthinbits is 16 (or None), then b2a_l() will generate a 4-character string.
|
||||
Note that b2a_l() does not mask off unused least-significant bits, so for example if os is
|
||||
2 bytes long and lengthinbits is 15, then you must ensure that the unused least-significant bit
|
||||
of os is a zero bit or you will get the wrong result. This precondition is tested by assertions
|
||||
if assertions are enabled.
|
||||
|
||||
Warning: if you generate a base-32 encoded string with b2a_l(), and then someone else tries to
|
||||
decode it by calling a2b() instead of a2b_l(), then they will (probably) get a different
|
||||
string than the one you encoded! So only use b2a_l() when you are sure that the encoding and
|
||||
decoding sides know exactly which lengthinbits to use. If you do not have a way for the
|
||||
encoder and the decoder to agree upon the lengthinbits, then it is best to use b2a() and
|
||||
a2b(). The only drawback to using b2a() over b2a_l() is that when you have a number of
|
||||
bits to encode that is not a multiple of 8, b2a() can sometimes generate a base-32 encoded
|
||||
string that is one or two characters longer than necessary.
|
||||
|
||||
@return the contents of os in base-32 encoded form
|
||||
"""
|
||||
precondition(isinstance(lengthinbits, (int, long,)), "lengthinbits is required to be an integer.", lengthinbits=lengthinbits)
|
||||
precondition((lengthinbits+7)/8 == len(os), "lengthinbits is required to specify a number of bits storable in exactly len(os) octets.", lengthinbits=lengthinbits, lenos=len(os))
|
||||
|
||||
os = map(ord, os)
|
||||
|
||||
numquintets = (lengthinbits+4)/5
|
||||
numoctetsofdata = (lengthinbits+7)/8
|
||||
# print "numoctetsofdata: %s, len(os): %s, lengthinbits: %s, numquintets: %s" % (numoctetsofdata, len(os), lengthinbits, numquintets,)
|
||||
# strip trailing octets that won't be used
|
||||
del os[numoctetsofdata:]
|
||||
# zero out any unused bits in the final octet
|
||||
if lengthinbits % 8 != 0:
|
||||
os[-1] = os[-1] >> (8-(lengthinbits % 8))
|
||||
os[-1] = os[-1] << (8-(lengthinbits % 8))
|
||||
# append zero octets for padding if needed
|
||||
numoctetsneeded = (numquintets*5+7)/8 + 1
|
||||
os.extend([0]*(numoctetsneeded-len(os)))
|
||||
|
||||
quintets = []
|
||||
cutoff = 256
|
||||
num = os[0]
|
||||
i = 0
|
||||
while len(quintets) < numquintets:
|
||||
i = i + 1
|
||||
assert len(os) > i, "len(os): %s, i: %s, len(quintets): %s, numquintets: %s, lengthinbits: %s, numoctetsofdata: %s, numoctetsneeded: %s, os: %s" % (len(os), i, len(quintets), numquintets, lengthinbits, numoctetsofdata, numoctetsneeded, os,)
|
||||
num = num * 256
|
||||
num = num + os[i]
|
||||
if cutoff == 1:
|
||||
cutoff = 256
|
||||
continue
|
||||
cutoff = cutoff * 8
|
||||
quintet = num / cutoff
|
||||
quintets.append(quintet)
|
||||
num = num - (quintet * cutoff)
|
||||
|
||||
cutoff = cutoff / 32
|
||||
quintet = num / cutoff
|
||||
quintets.append(quintet)
|
||||
num = num - (quintet * cutoff)
|
||||
|
||||
if len(quintets) > numquintets:
|
||||
assert len(quintets) == (numquintets+1), "len(quintets): %s, numquintets: %s, quintets: %s" % (len(quintets), numquintets, quintets,)
|
||||
quintets = quintets[:numquintets]
|
||||
res = string.translate(string.join(map(chr, quintets), ''), v2ctranstable)
|
||||
assert could_be_base32_encoded_l(res, lengthinbits), "lengthinbits: %s, res: %s" % (lengthinbits, res,)
|
||||
return res
|
||||
|
||||
# b2a() uses the minimal number of quintets sufficient to encode the binary
|
||||
# input. It just so happens that the relation is like this (everything is
|
||||
# modulo 40 bits).
|
||||
# num_qs = NUM_OS_TO_NUM_QS[num_os]
|
||||
NUM_OS_TO_NUM_QS=(0, 2, 4, 5, 7,)
|
||||
|
||||
# num_os = NUM_QS_TO_NUM_OS[num_qs], but if not NUM_QS_LEGIT[num_qs] then
|
||||
# there is *no* number of octets which would have resulted in this number of
|
||||
# quintets, so either the encoded string has been mangled (truncated) or else
|
||||
# you were supposed to decode it with a2b_l() (which means you were supposed
|
||||
# to know the actual length of the encoded data).
|
||||
|
||||
NUM_QS_TO_NUM_OS=(0, 1, 1, 2, 2, 3, 3, 4)
|
||||
NUM_QS_LEGIT=(1, 0, 1, 0, 1, 1, 0, 1,)
|
||||
NUM_QS_TO_NUM_BITS=tuple(map(lambda x: x*8, NUM_QS_TO_NUM_OS))
|
||||
|
||||
# A fast way to determine whether a given string *could* be base-32 encoded data, assuming that the
|
||||
# original data had 8K bits for a positive integer K.
|
||||
# The boolean value of s8[len(s)%8][ord(s[-1])], where s is the possibly base-32 encoded string
|
||||
# tells whether the final character is reasonable.
|
||||
def add_check_array(cs, sfmap):
|
||||
checka=[0] * 256
|
||||
for c in cs:
|
||||
checka[ord(c)] = 1
|
||||
sfmap.append(tuple(checka))
|
||||
|
||||
def init_s8():
|
||||
s8 = []
|
||||
add_check_array(chars, s8)
|
||||
for lenmod8 in (1, 2, 3, 4, 5, 6, 7,):
|
||||
if NUM_QS_LEGIT[lenmod8]:
|
||||
add_check_array(get_trailing_chars_without_lsbs(4-(NUM_QS_TO_NUM_BITS[lenmod8]%5)), s8)
|
||||
else:
|
||||
add_check_array('', s8)
|
||||
return tuple(s8)
|
||||
s8 = init_s8()
|
||||
|
||||
# A somewhat fast way to determine whether a given string *could* be base-32 encoded data, given a
|
||||
# lengthinbits.
|
||||
# The boolean value of s5[lengthinbits%5][ord(s[-1])], where s is the possibly base-32 encoded
|
||||
# string tells whether the final character is reasonable.
|
||||
def init_s5():
|
||||
s5 = []
|
||||
add_check_array(get_trailing_chars_without_lsbs(0), s5)
|
||||
for lenmod5 in [1,2,3,4]:
|
||||
add_check_array(get_trailing_chars_without_lsbs(5-lenmod5), s5)
|
||||
return tuple(s5)
|
||||
s5 = init_s5()
|
||||
|
||||
def could_be_base32_encoded(s, s8=s8, tr=string.translate, identitytranstable=identitytranstable, chars=chars):
|
||||
if s == '':
|
||||
return True
|
||||
return s8[len(s)%8][ord(s[-1])] and not tr(s, identitytranstable, chars)
|
||||
|
||||
def could_be_base32_encoded_l(s, lengthinbits, s5=s5, tr=string.translate, identitytranstable=identitytranstable, chars=chars):
|
||||
if s == '':
|
||||
return True
|
||||
assert lengthinbits%5 < len(s5), lengthinbits
|
||||
assert ord(s[-1]) < s5[lengthinbits%5]
|
||||
return (((lengthinbits+4)/5) == len(s)) and s5[lengthinbits%5][ord(s[-1])] and not string.translate(s, identitytranstable, chars)
|
||||
|
||||
def num_octets_that_encode_to_this_many_quintets(numqs):
|
||||
# Here is a computation that conveniently expresses this:
|
||||
return (numqs*5+3)/8
|
||||
|
||||
def a2b(cs):
|
||||
"""
|
||||
@param cs the base-32 encoded data (a string)
|
||||
"""
|
||||
precondition(could_be_base32_encoded(cs), "cs is required to be possibly base32 encoded data.", cs=cs)
|
||||
|
||||
return a2b_l(cs, num_octets_that_encode_to_this_many_quintets(len(cs))*8)
|
||||
|
||||
def a2b_l(cs, lengthinbits):
|
||||
"""
|
||||
@param lengthinbits the number of bits of data in encoded into cs
|
||||
|
||||
a2b_l() will return a result big enough to hold lengthinbits bits. So for example if cs is
|
||||
4 characters long (encoding at least 15 and up to 20 bits) and lengthinbits is 16, then a2b_l()
|
||||
will return a string of length 2 (since 2 bytes is sufficient to store 16 bits). If cs is 4
|
||||
characters long and lengthinbits is 20, then a2b_l() will return a string of length 3 (since
|
||||
3 bytes is sufficient to store 20 bits). Note that b2a_l() does not mask off unused least-
|
||||
significant bits, so for example if cs is 4 characters long and lengthinbits is 17, then you
|
||||
must ensure that all three of the unused least-significant bits of cs are zero bits or you will
|
||||
get the wrong result. This precondition is tested by assertions if assertions are enabled.
|
||||
(Generally you just require the encoder to ensure this consistency property between the least
|
||||
significant zero bits and value of lengthinbits, and reject strings that have a length-in-bits
|
||||
which isn't a multiple of 8 and yet don't have trailing zero bits, as improperly encoded.)
|
||||
|
||||
Please see the warning in the docstring of b2a_l() regarding the use of b2a() versus b2a_l().
|
||||
|
||||
@return the data encoded in cs
|
||||
"""
|
||||
precondition(could_be_base32_encoded_l(cs, lengthinbits), "cs is required to be possibly base32 encoded data.", cs=cs, lengthinbits=lengthinbits)
|
||||
if cs == '':
|
||||
return ''
|
||||
|
||||
qs = map(ord, string.translate(cs, c2vtranstable))
|
||||
|
||||
numoctets = (lengthinbits+7)/8
|
||||
numquintetsofdata = (lengthinbits+4)/5
|
||||
# strip trailing quintets that won't be used
|
||||
del qs[numquintetsofdata:]
|
||||
# zero out any unused bits in the final quintet
|
||||
if lengthinbits % 5 != 0:
|
||||
qs[-1] = qs[-1] >> (5-(lengthinbits % 5))
|
||||
qs[-1] = qs[-1] << (5-(lengthinbits % 5))
|
||||
# append zero quintets for padding if needed
|
||||
numquintetsneeded = (numoctets*8+4)/5
|
||||
qs.extend([0]*(numquintetsneeded-len(qs)))
|
||||
|
||||
octets = []
|
||||
pos = 2048
|
||||
num = qs[0] * pos
|
||||
readybits = 5
|
||||
i = 1
|
||||
while len(octets) < numoctets:
|
||||
while pos > 256:
|
||||
pos = pos / 32
|
||||
num = num + (qs[i] * pos)
|
||||
i = i + 1
|
||||
octet = num / 256
|
||||
octets.append(octet)
|
||||
num = num - (octet * 256)
|
||||
num = num * 256
|
||||
pos = pos * 256
|
||||
assert len(octets) == numoctets, "len(octets): %s, numoctets: %s, octets: %s" % (len(octets), numoctets, octets,)
|
||||
res = ''.join(map(chr, octets))
|
||||
precondition(b2a_l(res, lengthinbits) == cs, "cs is required to be the canonical base-32 encoding of some data.", b2a(res), res=res, cs=cs)
|
||||
return res
|
||||
|
||||
from foolscap import base32
|
||||
def nodeid_b2a(nodeid):
|
||||
|
@ -6,7 +6,7 @@ from twisted.internet import defer, address
|
||||
from twisted.internet.interfaces import IConsumer
|
||||
from nevow import inevow, rend, loaders, appserver, url, tags as T
|
||||
from nevow.static import File as nevow_File # TODO: merge with static.File?
|
||||
from allmydata.util import fileutil, idlib, observer, log
|
||||
from allmydata.util import base32, fileutil, idlib, observer, log
|
||||
import simplejson
|
||||
from allmydata.interfaces import IDownloadTarget, IDirectoryNode, IFileNode, \
|
||||
IMutableFileNode
|
||||
@ -1567,7 +1567,7 @@ class Status(rend.Page):
|
||||
|
||||
def _render_common(self, ctx, data):
|
||||
s = data
|
||||
si_s = idlib.b2a_or_none(s.get_storage_index())
|
||||
si_s = base32.b2a_or_none(s.get_storage_index())
|
||||
if si_s is None:
|
||||
si_s = "(None)"
|
||||
ctx.fillSlots("si", si_s)
|
||||
|
Loading…
Reference in New Issue
Block a user