636 lines
22 KiB
Python
636 lines
22 KiB
Python
"""Functions that read and write gzipped files.
|
|
|
|
The user of the file doesn't have to worry about the compression,
|
|
but random access is not allowed."""
|
|
|
|
# based on Andrew Kuchling's minigzip.py distributed with the zlib module
|
|
|
|
import struct, sys, time, os
|
|
import zlib
|
|
import builtins
|
|
import io
|
|
|
|
__all__ = ["GzipFile", "open", "compress", "decompress"]
|
|
|
|
FTEXT, FHCRC, FEXTRA, FNAME, FCOMMENT = 1, 2, 4, 8, 16
|
|
|
|
READ, WRITE = 1, 2
|
|
|
|
def write32u(output, value):
|
|
# The L format writes the bit pattern correctly whether signed
|
|
# or unsigned.
|
|
output.write(struct.pack("<L", value))
|
|
|
|
def read32(input):
|
|
return struct.unpack("<I", input.read(4))[0]
|
|
|
|
def open(filename, mode="rb", compresslevel=9):
|
|
"""Shorthand for GzipFile(filename, mode, compresslevel).
|
|
|
|
The filename argument is required; mode defaults to 'rb'
|
|
and compresslevel defaults to 9.
|
|
|
|
"""
|
|
return GzipFile(filename, mode, compresslevel)
|
|
|
|
class _PaddedFile:
|
|
"""Minimal read-only file object that prepends a string to the contents
|
|
of an actual file. Shouldn't be used outside of gzip.py, as it lacks
|
|
essential functionality."""
|
|
|
|
def __init__(self, f, prepend=b''):
|
|
self._buffer = prepend
|
|
self._length = len(prepend)
|
|
self.file = f
|
|
self._read = 0
|
|
|
|
def read(self, size):
|
|
if self._read is None:
|
|
return self.file.read(size)
|
|
if self._read + size <= self._length:
|
|
read = self._read
|
|
self._read += size
|
|
return self._buffer[read:self._read]
|
|
else:
|
|
read = self._read
|
|
self._read = None
|
|
return self._buffer[read:] + \
|
|
self.file.read(size-self._length+read)
|
|
|
|
def prepend(self, prepend=b'', readprevious=False):
|
|
if self._read is None:
|
|
self._buffer = prepend
|
|
elif readprevious and len(prepend) <= self._read:
|
|
self._read -= len(prepend)
|
|
return
|
|
else:
|
|
self._buffer = self._buffer[read:] + prepend
|
|
self._length = len(self._buffer)
|
|
self._read = 0
|
|
|
|
def unused(self):
|
|
if self._read is None:
|
|
return b''
|
|
return self._buffer[self._read:]
|
|
|
|
def seek(self, offset, whence=0):
|
|
# This is only ever called with offset=whence=0
|
|
if whence == 1 and self._read is not None:
|
|
if 0 <= offset + self._read <= self._length:
|
|
self._read += offset
|
|
return
|
|
else:
|
|
offset += self._length - self._read
|
|
self._read = None
|
|
self._buffer = None
|
|
return self.file.seek(offset, whence)
|
|
|
|
def __getattr__(self, name):
|
|
return getattr(self.file, name)
|
|
|
|
|
|
class GzipFile(io.BufferedIOBase):
|
|
"""The GzipFile class simulates most of the methods of a file object with
|
|
the exception of the readinto() and truncate() methods.
|
|
|
|
"""
|
|
|
|
myfileobj = None
|
|
max_read_chunk = 10 * 1024 * 1024 # 10Mb
|
|
|
|
def __init__(self, filename=None, mode=None,
|
|
compresslevel=9, fileobj=None, mtime=None):
|
|
"""Constructor for the GzipFile class.
|
|
|
|
At least one of fileobj and filename must be given a
|
|
non-trivial value.
|
|
|
|
The new class instance is based on fileobj, which can be a regular
|
|
file, a StringIO object, or any other object which simulates a file.
|
|
It defaults to None, in which case filename is opened to provide
|
|
a file object.
|
|
|
|
When fileobj is not None, the filename argument is only used to be
|
|
included in the gzip file header, which may includes the original
|
|
filename of the uncompressed file. It defaults to the filename of
|
|
fileobj, if discernible; otherwise, it defaults to the empty string,
|
|
and in this case the original filename is not included in the header.
|
|
|
|
The mode argument can be any of 'r', 'rb', 'a', 'ab', 'w', or 'wb',
|
|
depending on whether the file will be read or written. The default
|
|
is the mode of fileobj if discernible; otherwise, the default is 'rb'.
|
|
Be aware that only the 'rb', 'ab', and 'wb' values should be used
|
|
for cross-platform portability.
|
|
|
|
The compresslevel argument is an integer from 1 to 9 controlling the
|
|
level of compression; 1 is fastest and produces the least compression,
|
|
and 9 is slowest and produces the most compression. The default is 9.
|
|
|
|
The mtime argument is an optional numeric timestamp to be written
|
|
to the stream when compressing. All gzip compressed streams
|
|
are required to contain a timestamp. If omitted or None, the
|
|
current time is used. This module ignores the timestamp when
|
|
decompressing; however, some programs, such as gunzip, make use
|
|
of it. The format of the timestamp is the same as that of the
|
|
return value of time.time() and of the st_mtime member of the
|
|
object returned by os.stat().
|
|
|
|
"""
|
|
|
|
# guarantee the file is opened in binary mode on platforms
|
|
# that care about that sort of thing
|
|
if mode and 'b' not in mode:
|
|
mode += 'b'
|
|
if fileobj is None:
|
|
fileobj = self.myfileobj = builtins.open(filename, mode or 'rb')
|
|
if filename is None:
|
|
if hasattr(fileobj, 'name'): filename = fileobj.name
|
|
else: filename = ''
|
|
if mode is None:
|
|
if hasattr(fileobj, 'mode'): mode = fileobj.mode
|
|
else: mode = 'rb'
|
|
|
|
if mode[0:1] == 'r':
|
|
self.mode = READ
|
|
# Set flag indicating start of a new member
|
|
self._new_member = True
|
|
# Buffer data read from gzip file. extrastart is offset in
|
|
# stream where buffer starts. extrasize is number of
|
|
# bytes remaining in buffer from current stream position.
|
|
self.extrabuf = b""
|
|
self.extrasize = 0
|
|
self.extrastart = 0
|
|
self.name = filename
|
|
# Starts small, scales exponentially
|
|
self.min_readsize = 100
|
|
fileobj = _PaddedFile(fileobj)
|
|
|
|
elif mode[0:1] == 'w' or mode[0:1] == 'a':
|
|
self.mode = WRITE
|
|
self._init_write(filename)
|
|
self.compress = zlib.compressobj(compresslevel,
|
|
zlib.DEFLATED,
|
|
-zlib.MAX_WBITS,
|
|
zlib.DEF_MEM_LEVEL,
|
|
0)
|
|
else:
|
|
raise IOError("Mode " + mode + " not supported")
|
|
|
|
self.fileobj = fileobj
|
|
self.offset = 0
|
|
self.mtime = mtime
|
|
|
|
if self.mode == WRITE:
|
|
self._write_gzip_header()
|
|
|
|
@property
|
|
def filename(self):
|
|
import warnings
|
|
warnings.warn("use the name attribute", DeprecationWarning, 2)
|
|
if self.mode == WRITE and self.name[-3:] != ".gz":
|
|
return self.name + ".gz"
|
|
return self.name
|
|
|
|
def __repr__(self):
|
|
fileobj = self.fileobj
|
|
if isinstance(fileobj, _PaddedFile):
|
|
fileobj = fileobj.file
|
|
s = repr(fileobj)
|
|
return '<gzip ' + s[1:-1] + ' ' + hex(id(self)) + '>'
|
|
|
|
def _check_closed(self):
|
|
"""Raises a ValueError if the underlying file object has been closed.
|
|
|
|
"""
|
|
if self.closed:
|
|
raise ValueError('I/O operation on closed file.')
|
|
|
|
def _init_write(self, filename):
|
|
self.name = filename
|
|
self.crc = zlib.crc32(b"") & 0xffffffff
|
|
self.size = 0
|
|
self.writebuf = []
|
|
self.bufsize = 0
|
|
|
|
def _write_gzip_header(self):
|
|
self.fileobj.write(b'\037\213') # magic header
|
|
self.fileobj.write(b'\010') # compression method
|
|
try:
|
|
# RFC 1952 requires the FNAME field to be Latin-1. Do not
|
|
# include filenames that cannot be represented that way.
|
|
fname = os.path.basename(self.name)
|
|
fname = fname.encode('latin-1')
|
|
if fname.endswith(b'.gz'):
|
|
fname = fname[:-3]
|
|
except UnicodeEncodeError:
|
|
fname = b''
|
|
flags = 0
|
|
if fname:
|
|
flags = FNAME
|
|
self.fileobj.write(chr(flags).encode('latin-1'))
|
|
mtime = self.mtime
|
|
if mtime is None:
|
|
mtime = time.time()
|
|
write32u(self.fileobj, int(mtime))
|
|
self.fileobj.write(b'\002')
|
|
self.fileobj.write(b'\377')
|
|
if fname:
|
|
self.fileobj.write(fname + b'\000')
|
|
|
|
def _init_read(self):
|
|
self.crc = zlib.crc32(b"") & 0xffffffff
|
|
self.size = 0
|
|
|
|
def _read_gzip_header(self):
|
|
magic = self.fileobj.read(2)
|
|
if magic == b'':
|
|
raise EOFError("Reached EOF")
|
|
|
|
if magic != b'\037\213':
|
|
raise IOError('Not a gzipped file')
|
|
method = ord( self.fileobj.read(1) )
|
|
if method != 8:
|
|
raise IOError('Unknown compression method')
|
|
flag = ord( self.fileobj.read(1) )
|
|
self.mtime = read32(self.fileobj)
|
|
# extraflag = self.fileobj.read(1)
|
|
# os = self.fileobj.read(1)
|
|
self.fileobj.read(2)
|
|
|
|
if flag & FEXTRA:
|
|
# Read & discard the extra field, if present
|
|
xlen = ord(self.fileobj.read(1))
|
|
xlen = xlen + 256*ord(self.fileobj.read(1))
|
|
self.fileobj.read(xlen)
|
|
if flag & FNAME:
|
|
# Read and discard a null-terminated string containing the filename
|
|
while True:
|
|
s = self.fileobj.read(1)
|
|
if not s or s==b'\000':
|
|
break
|
|
if flag & FCOMMENT:
|
|
# Read and discard a null-terminated string containing a comment
|
|
while True:
|
|
s = self.fileobj.read(1)
|
|
if not s or s==b'\000':
|
|
break
|
|
if flag & FHCRC:
|
|
self.fileobj.read(2) # Read & discard the 16-bit header CRC
|
|
|
|
unused = self.fileobj.unused()
|
|
if unused:
|
|
uncompress = self.decompress.decompress(unused)
|
|
self._add_read_data(uncompress)
|
|
|
|
def write(self,data):
|
|
self._check_closed()
|
|
if self.mode != WRITE:
|
|
import errno
|
|
raise IOError(errno.EBADF, "write() on read-only GzipFile object")
|
|
|
|
if self.fileobj is None:
|
|
raise ValueError("write() on closed GzipFile object")
|
|
|
|
# Convert data type if called by io.BufferedWriter.
|
|
if isinstance(data, memoryview):
|
|
data = data.tobytes()
|
|
|
|
if len(data) > 0:
|
|
self.size = self.size + len(data)
|
|
self.crc = zlib.crc32(data, self.crc) & 0xffffffff
|
|
self.fileobj.write( self.compress.compress(data) )
|
|
self.offset += len(data)
|
|
|
|
return len(data)
|
|
|
|
def read(self, size=-1):
|
|
self._check_closed()
|
|
if self.mode != READ:
|
|
import errno
|
|
raise IOError(errno.EBADF, "read() on write-only GzipFile object")
|
|
|
|
if self.extrasize <= 0 and self.fileobj is None:
|
|
return b''
|
|
|
|
readsize = 1024
|
|
if size < 0: # get the whole thing
|
|
try:
|
|
while True:
|
|
self._read(readsize)
|
|
readsize = min(self.max_read_chunk, readsize * 2)
|
|
except EOFError:
|
|
size = self.extrasize
|
|
else: # just get some more of it
|
|
try:
|
|
while size > self.extrasize:
|
|
self._read(readsize)
|
|
readsize = min(self.max_read_chunk, readsize * 2)
|
|
except EOFError:
|
|
if size > self.extrasize:
|
|
size = self.extrasize
|
|
|
|
offset = self.offset - self.extrastart
|
|
chunk = self.extrabuf[offset: offset + size]
|
|
self.extrasize = self.extrasize - size
|
|
|
|
self.offset += size
|
|
return chunk
|
|
|
|
def read1(self, size=-1):
|
|
self._check_closed()
|
|
if self.mode != READ:
|
|
import errno
|
|
raise IOError(errno.EBADF, "read1() on write-only GzipFile object")
|
|
|
|
if self.extrasize <= 0 and self.fileobj is None:
|
|
return b''
|
|
|
|
try:
|
|
self._read()
|
|
except EOFError:
|
|
pass
|
|
if size < 0 or size > self.extrasize:
|
|
size = self.extrasize
|
|
|
|
offset = self.offset - self.extrastart
|
|
chunk = self.extrabuf[offset: offset + size]
|
|
self.extrasize -= size
|
|
self.offset += size
|
|
return chunk
|
|
|
|
def peek(self, n):
|
|
if self.mode != READ:
|
|
import errno
|
|
raise IOError(errno.EBADF, "peek() on write-only GzipFile object")
|
|
|
|
# Do not return ridiculously small buffers, for one common idiom
|
|
# is to call peek(1) and expect more bytes in return.
|
|
if n < 100:
|
|
n = 100
|
|
if self.extrasize == 0:
|
|
if self.fileobj is None:
|
|
return b''
|
|
try:
|
|
# 1024 is the same buffering heuristic used in read()
|
|
self._read(max(n, 1024))
|
|
except EOFError:
|
|
pass
|
|
offset = self.offset - self.extrastart
|
|
remaining = self.extrasize
|
|
assert remaining == len(self.extrabuf) - offset
|
|
return self.extrabuf[offset:offset + n]
|
|
|
|
def _unread(self, buf):
|
|
self.extrasize = len(buf) + self.extrasize
|
|
self.offset -= len(buf)
|
|
|
|
def _read(self, size=1024):
|
|
if self.fileobj is None:
|
|
raise EOFError("Reached EOF")
|
|
|
|
if self._new_member:
|
|
# If the _new_member flag is set, we have to
|
|
# jump to the next member, if there is one.
|
|
self._init_read()
|
|
self._read_gzip_header()
|
|
self.decompress = zlib.decompressobj(-zlib.MAX_WBITS)
|
|
self._new_member = False
|
|
|
|
# Read a chunk of data from the file
|
|
buf = self.fileobj.read(size)
|
|
|
|
# If the EOF has been reached, flush the decompression object
|
|
# and mark this object as finished.
|
|
|
|
if buf == b"":
|
|
uncompress = self.decompress.flush()
|
|
# Prepend the already read bytes to the fileobj to they can be
|
|
# seen by _read_eof()
|
|
self.fileobj.prepend(self.decompress.unused_data, True)
|
|
self._read_eof()
|
|
self._add_read_data( uncompress )
|
|
raise EOFError('Reached EOF')
|
|
|
|
uncompress = self.decompress.decompress(buf)
|
|
self._add_read_data( uncompress )
|
|
|
|
if self.decompress.unused_data != b"":
|
|
# Ending case: we've come to the end of a member in the file,
|
|
# so seek back to the start of the unused data, finish up
|
|
# this member, and read a new gzip header.
|
|
# Prepend the already read bytes to the fileobj to they can be
|
|
# seen by _read_eof() and _read_gzip_header()
|
|
self.fileobj.prepend(self.decompress.unused_data, True)
|
|
# Check the CRC and file size, and set the flag so we read
|
|
# a new member on the next call
|
|
self._read_eof()
|
|
self._new_member = True
|
|
|
|
def _add_read_data(self, data):
|
|
self.crc = zlib.crc32(data, self.crc) & 0xffffffff
|
|
offset = self.offset - self.extrastart
|
|
self.extrabuf = self.extrabuf[offset:] + data
|
|
self.extrasize = self.extrasize + len(data)
|
|
self.extrastart = self.offset
|
|
self.size = self.size + len(data)
|
|
|
|
def _read_eof(self):
|
|
# We've read to the end of the file
|
|
# We check the that the computed CRC and size of the
|
|
# uncompressed data matches the stored values. Note that the size
|
|
# stored is the true file size mod 2**32.
|
|
crc32 = read32(self.fileobj)
|
|
isize = read32(self.fileobj) # may exceed 2GB
|
|
if crc32 != self.crc:
|
|
raise IOError("CRC check failed %s != %s" % (hex(crc32),
|
|
hex(self.crc)))
|
|
elif isize != (self.size & 0xffffffff):
|
|
raise IOError("Incorrect length of data produced")
|
|
|
|
# Gzip files can be padded with zeroes and still have archives.
|
|
# Consume all zero bytes and set the file position to the first
|
|
# non-zero byte. See http://www.gzip.org/#faq8
|
|
c = b"\x00"
|
|
while c == b"\x00":
|
|
c = self.fileobj.read(1)
|
|
if c:
|
|
self.fileobj.prepend(c, True)
|
|
|
|
@property
|
|
def closed(self):
|
|
return self.fileobj is None
|
|
|
|
def close(self):
|
|
if self.fileobj is None:
|
|
return
|
|
if self.mode == WRITE:
|
|
self.fileobj.write(self.compress.flush())
|
|
write32u(self.fileobj, self.crc)
|
|
# self.size may exceed 2GB, or even 4GB
|
|
write32u(self.fileobj, self.size & 0xffffffff)
|
|
self.fileobj = None
|
|
elif self.mode == READ:
|
|
self.fileobj = None
|
|
if self.myfileobj:
|
|
self.myfileobj.close()
|
|
self.myfileobj = None
|
|
|
|
def flush(self,zlib_mode=zlib.Z_SYNC_FLUSH):
|
|
self._check_closed()
|
|
if self.mode == WRITE:
|
|
# Ensure the compressor's buffer is flushed
|
|
self.fileobj.write(self.compress.flush(zlib_mode))
|
|
self.fileobj.flush()
|
|
|
|
def fileno(self):
|
|
"""Invoke the underlying file object's fileno() method.
|
|
|
|
This will raise AttributeError if the underlying file object
|
|
doesn't support fileno().
|
|
"""
|
|
return self.fileobj.fileno()
|
|
|
|
def rewind(self):
|
|
'''Return the uncompressed stream file position indicator to the
|
|
beginning of the file'''
|
|
if self.mode != READ:
|
|
raise IOError("Can't rewind in write mode")
|
|
self.fileobj.seek(0)
|
|
self._new_member = True
|
|
self.extrabuf = b""
|
|
self.extrasize = 0
|
|
self.extrastart = 0
|
|
self.offset = 0
|
|
|
|
def readable(self):
|
|
return self.mode == READ
|
|
|
|
def writable(self):
|
|
return self.mode == WRITE
|
|
|
|
def seekable(self):
|
|
return True
|
|
|
|
def seek(self, offset, whence=0):
|
|
if whence:
|
|
if whence == 1:
|
|
offset = self.offset + offset
|
|
else:
|
|
raise ValueError('Seek from end not supported')
|
|
if self.mode == WRITE:
|
|
if offset < self.offset:
|
|
raise IOError('Negative seek in write mode')
|
|
count = offset - self.offset
|
|
chunk = bytes(1024)
|
|
for i in range(count // 1024):
|
|
self.write(chunk)
|
|
self.write(bytes(count % 1024))
|
|
elif self.mode == READ:
|
|
if offset < self.offset:
|
|
# for negative seek, rewind and do positive seek
|
|
self.rewind()
|
|
count = offset - self.offset
|
|
for i in range(count // 1024):
|
|
self.read(1024)
|
|
self.read(count % 1024)
|
|
|
|
return self.offset
|
|
|
|
def readline(self, size=-1):
|
|
if size < 0:
|
|
# Shortcut common case - newline found in buffer.
|
|
offset = self.offset - self.extrastart
|
|
i = self.extrabuf.find(b'\n', offset) + 1
|
|
if i > 0:
|
|
self.extrasize -= i - offset
|
|
self.offset += i - offset
|
|
return self.extrabuf[offset: i]
|
|
|
|
size = sys.maxsize
|
|
readsize = self.min_readsize
|
|
else:
|
|
readsize = size
|
|
bufs = []
|
|
while size != 0:
|
|
c = self.read(readsize)
|
|
i = c.find(b'\n')
|
|
|
|
# We set i=size to break out of the loop under two
|
|
# conditions: 1) there's no newline, and the chunk is
|
|
# larger than size, or 2) there is a newline, but the
|
|
# resulting line would be longer than 'size'.
|
|
if (size <= i) or (i == -1 and len(c) > size):
|
|
i = size - 1
|
|
|
|
if i >= 0 or c == b'':
|
|
bufs.append(c[:i + 1]) # Add portion of last chunk
|
|
self._unread(c[i + 1:]) # Push back rest of chunk
|
|
break
|
|
|
|
# Append chunk to list, decrease 'size',
|
|
bufs.append(c)
|
|
size = size - len(c)
|
|
readsize = min(size, readsize * 2)
|
|
if readsize > self.min_readsize:
|
|
self.min_readsize = min(readsize, self.min_readsize * 2, 512)
|
|
return b''.join(bufs) # Return resulting line
|
|
|
|
|
|
def compress(data, compresslevel=9):
|
|
"""Compress data in one shot and return the compressed string.
|
|
Optional argument is the compression level, in range of 1-9.
|
|
"""
|
|
buf = io.BytesIO()
|
|
with GzipFile(fileobj=buf, mode='wb', compresslevel=compresslevel) as f:
|
|
f.write(data)
|
|
return buf.getvalue()
|
|
|
|
def decompress(data):
|
|
"""Decompress a gzip compressed string in one shot.
|
|
Return the decompressed string.
|
|
"""
|
|
with GzipFile(fileobj=io.BytesIO(data)) as f:
|
|
return f.read()
|
|
|
|
|
|
def _test():
|
|
# Act like gzip; with -d, act like gunzip.
|
|
# The input file is not deleted, however, nor are any other gzip
|
|
# options or features supported.
|
|
args = sys.argv[1:]
|
|
decompress = args and args[0] == "-d"
|
|
if decompress:
|
|
args = args[1:]
|
|
if not args:
|
|
args = ["-"]
|
|
for arg in args:
|
|
if decompress:
|
|
if arg == "-":
|
|
f = GzipFile(filename="", mode="rb", fileobj=sys.stdin.buffer)
|
|
g = sys.stdout.buffer
|
|
else:
|
|
if arg[-3:] != ".gz":
|
|
print("filename doesn't end in .gz:", repr(arg))
|
|
continue
|
|
f = open(arg, "rb")
|
|
g = builtins.open(arg[:-3], "wb")
|
|
else:
|
|
if arg == "-":
|
|
f = sys.stdin.buffer
|
|
g = GzipFile(filename="", mode="wb", fileobj=sys.stdout.buffer)
|
|
else:
|
|
f = builtins.open(arg, "rb")
|
|
g = open(arg + ".gz", "wb")
|
|
while True:
|
|
chunk = f.read(1024)
|
|
if not chunk:
|
|
break
|
|
g.write(chunk)
|
|
if g is not sys.stdout:
|
|
g.close()
|
|
if f is not sys.stdin:
|
|
f.close()
|
|
|
|
if __name__ == '__main__':
|
|
_test()
|