|  | #!/usr/bin/env python3 | 
|  | #------------------------------------------------------------------- | 
|  | # tarfile.py | 
|  | #------------------------------------------------------------------- | 
|  | # Copyright (C) 2002 Lars Gustaebel <lars@gustaebel.de> | 
|  | # All rights reserved. | 
|  | # | 
|  | # Permission  is  hereby granted,  free  of charge,  to  any person | 
|  | # obtaining a  copy of  this software  and associated documentation | 
|  | # files  (the  "Software"),  to   deal  in  the  Software   without | 
|  | # restriction,  including  without limitation  the  rights to  use, | 
|  | # copy, modify, merge, publish, distribute, sublicense, and/or sell | 
|  | # copies  of  the  Software,  and to  permit  persons  to  whom the | 
|  | # Software  is  furnished  to  do  so,  subject  to  the  following | 
|  | # conditions: | 
|  | # | 
|  | # The above copyright  notice and this  permission notice shall  be | 
|  | # included in all copies or substantial portions of the Software. | 
|  | # | 
|  | # THE SOFTWARE IS PROVIDED "AS  IS", WITHOUT WARRANTY OF ANY  KIND, | 
|  | # EXPRESS OR IMPLIED, INCLUDING  BUT NOT LIMITED TO  THE WARRANTIES | 
|  | # OF  MERCHANTABILITY,  FITNESS   FOR  A  PARTICULAR   PURPOSE  AND | 
|  | # NONINFRINGEMENT.  IN  NO  EVENT SHALL  THE  AUTHORS  OR COPYRIGHT | 
|  | # HOLDERS  BE LIABLE  FOR ANY  CLAIM, DAMAGES  OR OTHER  LIABILITY, | 
|  | # WHETHER  IN AN  ACTION OF  CONTRACT, TORT  OR OTHERWISE,  ARISING | 
|  | # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | 
|  | # OTHER DEALINGS IN THE SOFTWARE. | 
|  | # | 
|  | """Read from and write to tar format archives. | 
|  | """ | 
|  |  | 
|  | version     = "0.9.0" | 
|  | __author__  = "Lars Gust\u00e4bel (lars@gustaebel.de)" | 
|  | __credits__ = "Gustavo Niemeyer, Niels Gust\u00e4bel, Richard Townsend." | 
|  |  | 
|  | #--------- | 
|  | # Imports | 
|  | #--------- | 
|  | from builtins import open as bltn_open | 
|  | import sys | 
|  | import os | 
|  | import io | 
|  | import shutil | 
|  | import stat | 
|  | import time | 
|  | import struct | 
|  | import copy | 
|  | import re | 
|  |  | 
|  | try: | 
|  | import pwd | 
|  | except ImportError: | 
|  | pwd = None | 
|  | try: | 
|  | import grp | 
|  | except ImportError: | 
|  | grp = None | 
|  |  | 
|  | # os.symlink on Windows prior to 6.0 raises NotImplementedError | 
|  | symlink_exception = (AttributeError, NotImplementedError) | 
|  | try: | 
|  | # OSError (winerror=1314) will be raised if the caller does not hold the | 
|  | # SeCreateSymbolicLinkPrivilege privilege | 
|  | symlink_exception += (OSError,) | 
|  | except NameError: | 
|  | pass | 
|  |  | 
|  | # from tarfile import * | 
|  | __all__ = ["TarFile", "TarInfo", "is_tarfile", "TarError", "ReadError", | 
|  | "CompressionError", "StreamError", "ExtractError", "HeaderError", | 
|  | "ENCODING", "USTAR_FORMAT", "GNU_FORMAT", "PAX_FORMAT", | 
|  | "DEFAULT_FORMAT", "open"] | 
|  |  | 
|  | #--------------------------------------------------------- | 
|  | # tar constants | 
|  | #--------------------------------------------------------- | 
|  | NUL = b"\0"                     # the null character | 
|  | BLOCKSIZE = 512                 # length of processing blocks | 
|  | RECORDSIZE = BLOCKSIZE * 20     # length of records | 
|  | GNU_MAGIC = b"ustar  \0"        # magic gnu tar string | 
|  | POSIX_MAGIC = b"ustar\x0000"    # magic posix tar string | 
|  |  | 
|  | LENGTH_NAME = 100               # maximum length of a filename | 
|  | LENGTH_LINK = 100               # maximum length of a linkname | 
|  | LENGTH_PREFIX = 155             # maximum length of the prefix field | 
|  |  | 
|  | REGTYPE = b"0"                  # regular file | 
|  | AREGTYPE = b"\0"                # regular file | 
|  | LNKTYPE = b"1"                  # link (inside tarfile) | 
|  | SYMTYPE = b"2"                  # symbolic link | 
|  | CHRTYPE = b"3"                  # character special device | 
|  | BLKTYPE = b"4"                  # block special device | 
|  | DIRTYPE = b"5"                  # directory | 
|  | FIFOTYPE = b"6"                 # fifo special device | 
|  | CONTTYPE = b"7"                 # contiguous file | 
|  |  | 
|  | GNUTYPE_LONGNAME = b"L"         # GNU tar longname | 
|  | GNUTYPE_LONGLINK = b"K"         # GNU tar longlink | 
|  | GNUTYPE_SPARSE = b"S"           # GNU tar sparse file | 
|  |  | 
|  | XHDTYPE = b"x"                  # POSIX.1-2001 extended header | 
|  | XGLTYPE = b"g"                  # POSIX.1-2001 global header | 
|  | SOLARIS_XHDTYPE = b"X"          # Solaris extended header | 
|  |  | 
|  | USTAR_FORMAT = 0                # POSIX.1-1988 (ustar) format | 
|  | GNU_FORMAT = 1                  # GNU tar format | 
|  | PAX_FORMAT = 2                  # POSIX.1-2001 (pax) format | 
|  | DEFAULT_FORMAT = PAX_FORMAT | 
|  |  | 
|  | #--------------------------------------------------------- | 
|  | # tarfile constants | 
|  | #--------------------------------------------------------- | 
|  | # File types that tarfile supports: | 
|  | SUPPORTED_TYPES = (REGTYPE, AREGTYPE, LNKTYPE, | 
|  | SYMTYPE, DIRTYPE, FIFOTYPE, | 
|  | CONTTYPE, CHRTYPE, BLKTYPE, | 
|  | GNUTYPE_LONGNAME, GNUTYPE_LONGLINK, | 
|  | GNUTYPE_SPARSE) | 
|  |  | 
|  | # File types that will be treated as a regular file. | 
|  | REGULAR_TYPES = (REGTYPE, AREGTYPE, | 
|  | CONTTYPE, GNUTYPE_SPARSE) | 
|  |  | 
|  | # File types that are part of the GNU tar format. | 
|  | GNU_TYPES = (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK, | 
|  | GNUTYPE_SPARSE) | 
|  |  | 
|  | # Fields from a pax header that override a TarInfo attribute. | 
|  | PAX_FIELDS = ("path", "linkpath", "size", "mtime", | 
|  | "uid", "gid", "uname", "gname") | 
|  |  | 
|  | # Fields from a pax header that are affected by hdrcharset. | 
|  | PAX_NAME_FIELDS = {"path", "linkpath", "uname", "gname"} | 
|  |  | 
|  | # Fields in a pax header that are numbers, all other fields | 
|  | # are treated as strings. | 
|  | PAX_NUMBER_FIELDS = { | 
|  | "atime": float, | 
|  | "ctime": float, | 
|  | "mtime": float, | 
|  | "uid": int, | 
|  | "gid": int, | 
|  | "size": int | 
|  | } | 
|  |  | 
|  | #--------------------------------------------------------- | 
|  | # initialization | 
|  | #--------------------------------------------------------- | 
|  | if os.name == "nt": | 
|  | ENCODING = "utf-8" | 
|  | else: | 
|  | ENCODING = sys.getfilesystemencoding() | 
|  |  | 
|  | #--------------------------------------------------------- | 
|  | # Some useful functions | 
|  | #--------------------------------------------------------- | 
|  |  | 
|  | def stn(s, length, encoding, errors): | 
|  | """Convert a string to a null-terminated bytes object. | 
|  | """ | 
|  | s = s.encode(encoding, errors) | 
|  | return s[:length] + (length - len(s)) * NUL | 
|  |  | 
|  | def nts(s, encoding, errors): | 
|  | """Convert a null-terminated bytes object to a string. | 
|  | """ | 
|  | p = s.find(b"\0") | 
|  | if p != -1: | 
|  | s = s[:p] | 
|  | return s.decode(encoding, errors) | 
|  |  | 
|  | def nti(s): | 
|  | """Convert a number field to a python number. | 
|  | """ | 
|  | # There are two possible encodings for a number field, see | 
|  | # itn() below. | 
|  | if s[0] in (0o200, 0o377): | 
|  | n = 0 | 
|  | for i in range(len(s) - 1): | 
|  | n <<= 8 | 
|  | n += s[i + 1] | 
|  | if s[0] == 0o377: | 
|  | n = -(256 ** (len(s) - 1) - n) | 
|  | else: | 
|  | try: | 
|  | s = nts(s, "ascii", "strict") | 
|  | n = int(s.strip() or "0", 8) | 
|  | except ValueError: | 
|  | raise InvalidHeaderError("invalid header") | 
|  | return n | 
|  |  | 
|  | def itn(n, digits=8, format=DEFAULT_FORMAT): | 
|  | """Convert a python number to a number field. | 
|  | """ | 
|  | # POSIX 1003.1-1988 requires numbers to be encoded as a string of | 
|  | # octal digits followed by a null-byte, this allows values up to | 
|  | # (8**(digits-1))-1. GNU tar allows storing numbers greater than | 
|  | # that if necessary. A leading 0o200 or 0o377 byte indicate this | 
|  | # particular encoding, the following digits-1 bytes are a big-endian | 
|  | # base-256 representation. This allows values up to (256**(digits-1))-1. | 
|  | # A 0o200 byte indicates a positive number, a 0o377 byte a negative | 
|  | # number. | 
|  | n = int(n) | 
|  | if 0 <= n < 8 ** (digits - 1): | 
|  | s = bytes("%0*o" % (digits - 1, n), "ascii") + NUL | 
|  | elif format == GNU_FORMAT and -256 ** (digits - 1) <= n < 256 ** (digits - 1): | 
|  | if n >= 0: | 
|  | s = bytearray([0o200]) | 
|  | else: | 
|  | s = bytearray([0o377]) | 
|  | n = 256 ** digits + n | 
|  |  | 
|  | for i in range(digits - 1): | 
|  | s.insert(1, n & 0o377) | 
|  | n >>= 8 | 
|  | else: | 
|  | raise ValueError("overflow in number field") | 
|  |  | 
|  | return s | 
|  |  | 
|  | def calc_chksums(buf): | 
|  | """Calculate the checksum for a member's header by summing up all | 
|  | characters except for the chksum field which is treated as if | 
|  | it was filled with spaces. According to the GNU tar sources, | 
|  | some tars (Sun and NeXT) calculate chksum with signed char, | 
|  | which will be different if there are chars in the buffer with | 
|  | the high bit set. So we calculate two checksums, unsigned and | 
|  | signed. | 
|  | """ | 
|  | unsigned_chksum = 256 + sum(struct.unpack_from("148B8x356B", buf)) | 
|  | signed_chksum = 256 + sum(struct.unpack_from("148b8x356b", buf)) | 
|  | return unsigned_chksum, signed_chksum | 
|  |  | 
|  | def copyfileobj(src, dst, length=None, exception=OSError, bufsize=None): | 
|  | """Copy length bytes from fileobj src to fileobj dst. | 
|  | If length is None, copy the entire content. | 
|  | """ | 
|  | bufsize = bufsize or 16 * 1024 | 
|  | if length == 0: | 
|  | return | 
|  | if length is None: | 
|  | shutil.copyfileobj(src, dst, bufsize) | 
|  | return | 
|  |  | 
|  | blocks, remainder = divmod(length, bufsize) | 
|  | for b in range(blocks): | 
|  | buf = src.read(bufsize) | 
|  | if len(buf) < bufsize: | 
|  | raise exception("unexpected end of data") | 
|  | dst.write(buf) | 
|  |  | 
|  | if remainder != 0: | 
|  | buf = src.read(remainder) | 
|  | if len(buf) < remainder: | 
|  | raise exception("unexpected end of data") | 
|  | dst.write(buf) | 
|  | return | 
|  |  | 
|  | def _safe_print(s): | 
|  | encoding = getattr(sys.stdout, 'encoding', None) | 
|  | if encoding is not None: | 
|  | s = s.encode(encoding, 'backslashreplace').decode(encoding) | 
|  | print(s, end=' ') | 
|  |  | 
|  |  | 
|  | class TarError(Exception): | 
|  | """Base exception.""" | 
|  | pass | 
|  | class ExtractError(TarError): | 
|  | """General exception for extract errors.""" | 
|  | pass | 
|  | class ReadError(TarError): | 
|  | """Exception for unreadable tar archives.""" | 
|  | pass | 
|  | class CompressionError(TarError): | 
|  | """Exception for unavailable compression methods.""" | 
|  | pass | 
|  | class StreamError(TarError): | 
|  | """Exception for unsupported operations on stream-like TarFiles.""" | 
|  | pass | 
|  | class HeaderError(TarError): | 
|  | """Base exception for header errors.""" | 
|  | pass | 
|  | class EmptyHeaderError(HeaderError): | 
|  | """Exception for empty headers.""" | 
|  | pass | 
|  | class TruncatedHeaderError(HeaderError): | 
|  | """Exception for truncated headers.""" | 
|  | pass | 
|  | class EOFHeaderError(HeaderError): | 
|  | """Exception for end of file headers.""" | 
|  | pass | 
|  | class InvalidHeaderError(HeaderError): | 
|  | """Exception for invalid headers.""" | 
|  | pass | 
|  | class SubsequentHeaderError(HeaderError): | 
|  | """Exception for missing and invalid extended headers.""" | 
|  | pass | 
|  |  | 
|  | #--------------------------- | 
|  | # internal stream interface | 
|  | #--------------------------- | 
|  | class _LowLevelFile: | 
|  | """Low-level file object. Supports reading and writing. | 
|  | It is used instead of a regular file object for streaming | 
|  | access. | 
|  | """ | 
|  |  | 
|  | def __init__(self, name, mode): | 
|  | mode = { | 
|  | "r": os.O_RDONLY, | 
|  | "w": os.O_WRONLY | os.O_CREAT | os.O_TRUNC, | 
|  | }[mode] | 
|  | if hasattr(os, "O_BINARY"): | 
|  | mode |= os.O_BINARY | 
|  | self.fd = os.open(name, mode, 0o666) | 
|  |  | 
|  | def close(self): | 
|  | os.close(self.fd) | 
|  |  | 
|  | def read(self, size): | 
|  | return os.read(self.fd, size) | 
|  |  | 
|  | def write(self, s): | 
|  | os.write(self.fd, s) | 
|  |  | 
|  | class _Stream: | 
|  | """Class that serves as an adapter between TarFile and | 
|  | a stream-like object.  The stream-like object only | 
|  | needs to have a read() or write() method and is accessed | 
|  | blockwise.  Use of gzip or bzip2 compression is possible. | 
|  | A stream-like object could be for example: sys.stdin, | 
|  | sys.stdout, a socket, a tape device etc. | 
|  |  | 
|  | _Stream is intended to be used only internally. | 
|  | """ | 
|  |  | 
|  | def __init__(self, name, mode, comptype, fileobj, bufsize): | 
|  | """Construct a _Stream object. | 
|  | """ | 
|  | self._extfileobj = True | 
|  | if fileobj is None: | 
|  | fileobj = _LowLevelFile(name, mode) | 
|  | self._extfileobj = False | 
|  |  | 
|  | if comptype == '*': | 
|  | # Enable transparent compression detection for the | 
|  | # stream interface | 
|  | fileobj = _StreamProxy(fileobj) | 
|  | comptype = fileobj.getcomptype() | 
|  |  | 
|  | self.name     = name or "" | 
|  | self.mode     = mode | 
|  | self.comptype = comptype | 
|  | self.fileobj  = fileobj | 
|  | self.bufsize  = bufsize | 
|  | self.buf      = b"" | 
|  | self.pos      = 0 | 
|  | self.closed   = False | 
|  |  | 
|  | try: | 
|  | if comptype == "gz": | 
|  | try: | 
|  | import zlib | 
|  | except ImportError: | 
|  | raise CompressionError("zlib module is not available") | 
|  | self.zlib = zlib | 
|  | self.crc = zlib.crc32(b"") | 
|  | if mode == "r": | 
|  | self._init_read_gz() | 
|  | self.exception = zlib.error | 
|  | else: | 
|  | self._init_write_gz() | 
|  |  | 
|  | elif comptype == "bz2": | 
|  | try: | 
|  | import bz2 | 
|  | except ImportError: | 
|  | raise CompressionError("bz2 module is not available") | 
|  | if mode == "r": | 
|  | self.dbuf = b"" | 
|  | self.cmp = bz2.BZ2Decompressor() | 
|  | self.exception = OSError | 
|  | else: | 
|  | self.cmp = bz2.BZ2Compressor() | 
|  |  | 
|  | elif comptype == "xz": | 
|  | try: | 
|  | import lzma | 
|  | except ImportError: | 
|  | raise CompressionError("lzma module is not available") | 
|  | if mode == "r": | 
|  | self.dbuf = b"" | 
|  | self.cmp = lzma.LZMADecompressor() | 
|  | self.exception = lzma.LZMAError | 
|  | else: | 
|  | self.cmp = lzma.LZMACompressor() | 
|  |  | 
|  | elif comptype != "tar": | 
|  | raise CompressionError("unknown compression type %r" % comptype) | 
|  |  | 
|  | except: | 
|  | if not self._extfileobj: | 
|  | self.fileobj.close() | 
|  | self.closed = True | 
|  | raise | 
|  |  | 
|  | def __del__(self): | 
|  | if hasattr(self, "closed") and not self.closed: | 
|  | self.close() | 
|  |  | 
|  | def _init_write_gz(self): | 
|  | """Initialize for writing with gzip compression. | 
|  | """ | 
|  | self.cmp = self.zlib.compressobj(9, self.zlib.DEFLATED, | 
|  | -self.zlib.MAX_WBITS, | 
|  | self.zlib.DEF_MEM_LEVEL, | 
|  | 0) | 
|  | timestamp = struct.pack("<L", int(time.time())) | 
|  | self.__write(b"\037\213\010\010" + timestamp + b"\002\377") | 
|  | if self.name.endswith(".gz"): | 
|  | self.name = self.name[:-3] | 
|  | # Honor "directory components removed" from RFC1952 | 
|  | self.name = os.path.basename(self.name) | 
|  | # RFC1952 says we must use ISO-8859-1 for the FNAME field. | 
|  | self.__write(self.name.encode("iso-8859-1", "replace") + NUL) | 
|  |  | 
|  | def write(self, s): | 
|  | """Write string s to the stream. | 
|  | """ | 
|  | if self.comptype == "gz": | 
|  | self.crc = self.zlib.crc32(s, self.crc) | 
|  | self.pos += len(s) | 
|  | if self.comptype != "tar": | 
|  | s = self.cmp.compress(s) | 
|  | self.__write(s) | 
|  |  | 
|  | def __write(self, s): | 
|  | """Write string s to the stream if a whole new block | 
|  | is ready to be written. | 
|  | """ | 
|  | self.buf += s | 
|  | while len(self.buf) > self.bufsize: | 
|  | self.fileobj.write(self.buf[:self.bufsize]) | 
|  | self.buf = self.buf[self.bufsize:] | 
|  |  | 
|  | def close(self): | 
|  | """Close the _Stream object. No operation should be | 
|  | done on it afterwards. | 
|  | """ | 
|  | if self.closed: | 
|  | return | 
|  |  | 
|  | self.closed = True | 
|  | try: | 
|  | if self.mode == "w" and self.comptype != "tar": | 
|  | self.buf += self.cmp.flush() | 
|  |  | 
|  | if self.mode == "w" and self.buf: | 
|  | self.fileobj.write(self.buf) | 
|  | self.buf = b"" | 
|  | if self.comptype == "gz": | 
|  | self.fileobj.write(struct.pack("<L", self.crc)) | 
|  | self.fileobj.write(struct.pack("<L", self.pos & 0xffffFFFF)) | 
|  | finally: | 
|  | if not self._extfileobj: | 
|  | self.fileobj.close() | 
|  |  | 
|  | def _init_read_gz(self): | 
|  | """Initialize for reading a gzip compressed fileobj. | 
|  | """ | 
|  | self.cmp = self.zlib.decompressobj(-self.zlib.MAX_WBITS) | 
|  | self.dbuf = b"" | 
|  |  | 
|  | # taken from gzip.GzipFile with some alterations | 
|  | if self.__read(2) != b"\037\213": | 
|  | raise ReadError("not a gzip file") | 
|  | if self.__read(1) != b"\010": | 
|  | raise CompressionError("unsupported compression method") | 
|  |  | 
|  | flag = ord(self.__read(1)) | 
|  | self.__read(6) | 
|  |  | 
|  | if flag & 4: | 
|  | xlen = ord(self.__read(1)) + 256 * ord(self.__read(1)) | 
|  | self.read(xlen) | 
|  | if flag & 8: | 
|  | while True: | 
|  | s = self.__read(1) | 
|  | if not s or s == NUL: | 
|  | break | 
|  | if flag & 16: | 
|  | while True: | 
|  | s = self.__read(1) | 
|  | if not s or s == NUL: | 
|  | break | 
|  | if flag & 2: | 
|  | self.__read(2) | 
|  |  | 
|  | def tell(self): | 
|  | """Return the stream's file pointer position. | 
|  | """ | 
|  | return self.pos | 
|  |  | 
|  | def seek(self, pos=0): | 
|  | """Set the stream's file pointer to pos. Negative seeking | 
|  | is forbidden. | 
|  | """ | 
|  | if pos - self.pos >= 0: | 
|  | blocks, remainder = divmod(pos - self.pos, self.bufsize) | 
|  | for i in range(blocks): | 
|  | self.read(self.bufsize) | 
|  | self.read(remainder) | 
|  | else: | 
|  | raise StreamError("seeking backwards is not allowed") | 
|  | return self.pos | 
|  |  | 
|  | def read(self, size): | 
|  | """Return the next size number of bytes from the stream.""" | 
|  | assert size is not None | 
|  | buf = self._read(size) | 
|  | self.pos += len(buf) | 
|  | return buf | 
|  |  | 
|  | def _read(self, size): | 
|  | """Return size bytes from the stream. | 
|  | """ | 
|  | if self.comptype == "tar": | 
|  | return self.__read(size) | 
|  |  | 
|  | c = len(self.dbuf) | 
|  | t = [self.dbuf] | 
|  | while c < size: | 
|  | # Skip underlying buffer to avoid unaligned double buffering. | 
|  | if self.buf: | 
|  | buf = self.buf | 
|  | self.buf = b"" | 
|  | else: | 
|  | buf = self.fileobj.read(self.bufsize) | 
|  | if not buf: | 
|  | break | 
|  | try: | 
|  | buf = self.cmp.decompress(buf) | 
|  | except self.exception: | 
|  | raise ReadError("invalid compressed data") | 
|  | t.append(buf) | 
|  | c += len(buf) | 
|  | t = b"".join(t) | 
|  | self.dbuf = t[size:] | 
|  | return t[:size] | 
|  |  | 
|  | def __read(self, size): | 
|  | """Return size bytes from stream. If internal buffer is empty, | 
|  | read another block from the stream. | 
|  | """ | 
|  | c = len(self.buf) | 
|  | t = [self.buf] | 
|  | while c < size: | 
|  | buf = self.fileobj.read(self.bufsize) | 
|  | if not buf: | 
|  | break | 
|  | t.append(buf) | 
|  | c += len(buf) | 
|  | t = b"".join(t) | 
|  | self.buf = t[size:] | 
|  | return t[:size] | 
|  | # class _Stream | 
|  |  | 
|  | class _StreamProxy(object): | 
|  | """Small proxy class that enables transparent compression | 
|  | detection for the Stream interface (mode 'r|*'). | 
|  | """ | 
|  |  | 
|  | def __init__(self, fileobj): | 
|  | self.fileobj = fileobj | 
|  | self.buf = self.fileobj.read(BLOCKSIZE) | 
|  |  | 
|  | def read(self, size): | 
|  | self.read = self.fileobj.read | 
|  | return self.buf | 
|  |  | 
|  | def getcomptype(self): | 
|  | if self.buf.startswith(b"\x1f\x8b\x08"): | 
|  | return "gz" | 
|  | elif self.buf[0:3] == b"BZh" and self.buf[4:10] == b"1AY&SY": | 
|  | return "bz2" | 
|  | elif self.buf.startswith((b"\x5d\x00\x00\x80", b"\xfd7zXZ")): | 
|  | return "xz" | 
|  | else: | 
|  | return "tar" | 
|  |  | 
|  | def close(self): | 
|  | self.fileobj.close() | 
|  | # class StreamProxy | 
|  |  | 
|  | #------------------------ | 
|  | # Extraction file object | 
|  | #------------------------ | 
|  | class _FileInFile(object): | 
|  | """A thin wrapper around an existing file object that | 
|  | provides a part of its data as an individual file | 
|  | object. | 
|  | """ | 
|  |  | 
|  | def __init__(self, fileobj, offset, size, blockinfo=None): | 
|  | self.fileobj = fileobj | 
|  | self.offset = offset | 
|  | self.size = size | 
|  | self.position = 0 | 
|  | self.name = getattr(fileobj, "name", None) | 
|  | self.closed = False | 
|  |  | 
|  | if blockinfo is None: | 
|  | blockinfo = [(0, size)] | 
|  |  | 
|  | # Construct a map with data and zero blocks. | 
|  | self.map_index = 0 | 
|  | self.map = [] | 
|  | lastpos = 0 | 
|  | realpos = self.offset | 
|  | for offset, size in blockinfo: | 
|  | if offset > lastpos: | 
|  | self.map.append((False, lastpos, offset, None)) | 
|  | self.map.append((True, offset, offset + size, realpos)) | 
|  | realpos += size | 
|  | lastpos = offset + size | 
|  | if lastpos < self.size: | 
|  | self.map.append((False, lastpos, self.size, None)) | 
|  |  | 
|  | def flush(self): | 
|  | pass | 
|  |  | 
|  | def readable(self): | 
|  | return True | 
|  |  | 
|  | def writable(self): | 
|  | return False | 
|  |  | 
|  | def seekable(self): | 
|  | return self.fileobj.seekable() | 
|  |  | 
|  | def tell(self): | 
|  | """Return the current file position. | 
|  | """ | 
|  | return self.position | 
|  |  | 
|  | def seek(self, position, whence=io.SEEK_SET): | 
|  | """Seek to a position in the file. | 
|  | """ | 
|  | if whence == io.SEEK_SET: | 
|  | self.position = min(max(position, 0), self.size) | 
|  | elif whence == io.SEEK_CUR: | 
|  | if position < 0: | 
|  | self.position = max(self.position + position, 0) | 
|  | else: | 
|  | self.position = min(self.position + position, self.size) | 
|  | elif whence == io.SEEK_END: | 
|  | self.position = max(min(self.size + position, self.size), 0) | 
|  | else: | 
|  | raise ValueError("Invalid argument") | 
|  | return self.position | 
|  |  | 
|  | def read(self, size=None): | 
|  | """Read data from the file. | 
|  | """ | 
|  | if size is None: | 
|  | size = self.size - self.position | 
|  | else: | 
|  | size = min(size, self.size - self.position) | 
|  |  | 
|  | buf = b"" | 
|  | while size > 0: | 
|  | while True: | 
|  | data, start, stop, offset = self.map[self.map_index] | 
|  | if start <= self.position < stop: | 
|  | break | 
|  | else: | 
|  | self.map_index += 1 | 
|  | if self.map_index == len(self.map): | 
|  | self.map_index = 0 | 
|  | length = min(size, stop - self.position) | 
|  | if data: | 
|  | self.fileobj.seek(offset + (self.position - start)) | 
|  | b = self.fileobj.read(length) | 
|  | if len(b) != length: | 
|  | raise ReadError("unexpected end of data") | 
|  | buf += b | 
|  | else: | 
|  | buf += NUL * length | 
|  | size -= length | 
|  | self.position += length | 
|  | return buf | 
|  |  | 
|  | def readinto(self, b): | 
|  | buf = self.read(len(b)) | 
|  | b[:len(buf)] = buf | 
|  | return len(buf) | 
|  |  | 
|  | def close(self): | 
|  | self.closed = True | 
|  | #class _FileInFile | 
|  |  | 
|  | class ExFileObject(io.BufferedReader): | 
|  |  | 
|  | def __init__(self, tarfile, tarinfo): | 
|  | fileobj = _FileInFile(tarfile.fileobj, tarinfo.offset_data, | 
|  | tarinfo.size, tarinfo.sparse) | 
|  | super().__init__(fileobj) | 
|  | #class ExFileObject | 
|  |  | 
|  | #------------------ | 
|  | # Exported Classes | 
|  | #------------------ | 
|  | class TarInfo(object): | 
|  | """Informational class which holds the details about an | 
|  | archive member given by a tar header block. | 
|  | TarInfo objects are returned by TarFile.getmember(), | 
|  | TarFile.getmembers() and TarFile.gettarinfo() and are | 
|  | usually created internally. | 
|  | """ | 
|  |  | 
|  | __slots__ = dict( | 
|  | name = 'Name of the archive member.', | 
|  | mode = 'Permission bits.', | 
|  | uid = 'User ID of the user who originally stored this member.', | 
|  | gid = 'Group ID of the user who originally stored this member.', | 
|  | size = 'Size in bytes.', | 
|  | mtime = 'Time of last modification.', | 
|  | chksum = 'Header checksum.', | 
|  | type = ('File type. type is usually one of these constants: ' | 
|  | 'REGTYPE, AREGTYPE, LNKTYPE, SYMTYPE, DIRTYPE, FIFOTYPE, ' | 
|  | 'CONTTYPE, CHRTYPE, BLKTYPE, GNUTYPE_SPARSE.'), | 
|  | linkname = ('Name of the target file name, which is only present ' | 
|  | 'in TarInfo objects of type LNKTYPE and SYMTYPE.'), | 
|  | uname = 'User name.', | 
|  | gname = 'Group name.', | 
|  | devmajor = 'Device major number.', | 
|  | devminor = 'Device minor number.', | 
|  | offset = 'The tar header starts here.', | 
|  | offset_data = "The file's data starts here.", | 
|  | pax_headers = ('A dictionary containing key-value pairs of an ' | 
|  | 'associated pax extended header.'), | 
|  | sparse = 'Sparse member information.', | 
|  | tarfile = None, | 
|  | _sparse_structs = None, | 
|  | _link_target = None, | 
|  | ) | 
|  |  | 
|  | def __init__(self, name=""): | 
|  | """Construct a TarInfo object. name is the optional name | 
|  | of the member. | 
|  | """ | 
|  | self.name = name        # member name | 
|  | self.mode = 0o644       # file permissions | 
|  | self.uid = 0            # user id | 
|  | self.gid = 0            # group id | 
|  | self.size = 0           # file size | 
|  | self.mtime = 0          # modification time | 
|  | self.chksum = 0         # header checksum | 
|  | self.type = REGTYPE     # member type | 
|  | self.linkname = ""      # link name | 
|  | self.uname = ""         # user name | 
|  | self.gname = ""         # group name | 
|  | self.devmajor = 0       # device major number | 
|  | self.devminor = 0       # device minor number | 
|  |  | 
|  | self.offset = 0         # the tar header starts here | 
|  | self.offset_data = 0    # the file's data starts here | 
|  |  | 
|  | self.sparse = None      # sparse member information | 
|  | self.pax_headers = {}   # pax header information | 
|  |  | 
|  | @property | 
|  | def path(self): | 
|  | 'In pax headers, "name" is called "path".' | 
|  | return self.name | 
|  |  | 
|  | @path.setter | 
|  | def path(self, name): | 
|  | self.name = name | 
|  |  | 
|  | @property | 
|  | def linkpath(self): | 
|  | 'In pax headers, "linkname" is called "linkpath".' | 
|  | return self.linkname | 
|  |  | 
|  | @linkpath.setter | 
|  | def linkpath(self, linkname): | 
|  | self.linkname = linkname | 
|  |  | 
|  | def __repr__(self): | 
|  | return "<%s %r at %#x>" % (self.__class__.__name__,self.name,id(self)) | 
|  |  | 
|  | def get_info(self): | 
|  | """Return the TarInfo's attributes as a dictionary. | 
|  | """ | 
|  | info = { | 
|  | "name":     self.name, | 
|  | "mode":     self.mode & 0o7777, | 
|  | "uid":      self.uid, | 
|  | "gid":      self.gid, | 
|  | "size":     self.size, | 
|  | "mtime":    self.mtime, | 
|  | "chksum":   self.chksum, | 
|  | "type":     self.type, | 
|  | "linkname": self.linkname, | 
|  | "uname":    self.uname, | 
|  | "gname":    self.gname, | 
|  | "devmajor": self.devmajor, | 
|  | "devminor": self.devminor | 
|  | } | 
|  |  | 
|  | if info["type"] == DIRTYPE and not info["name"].endswith("/"): | 
|  | info["name"] += "/" | 
|  |  | 
|  | return info | 
|  |  | 
|  | def tobuf(self, format=DEFAULT_FORMAT, encoding=ENCODING, errors="surrogateescape"): | 
|  | """Return a tar header as a string of 512 byte blocks. | 
|  | """ | 
|  | info = self.get_info() | 
|  |  | 
|  | if format == USTAR_FORMAT: | 
|  | return self.create_ustar_header(info, encoding, errors) | 
|  | elif format == GNU_FORMAT: | 
|  | return self.create_gnu_header(info, encoding, errors) | 
|  | elif format == PAX_FORMAT: | 
|  | return self.create_pax_header(info, encoding) | 
|  | else: | 
|  | raise ValueError("invalid format") | 
|  |  | 
|  | def create_ustar_header(self, info, encoding, errors): | 
|  | """Return the object as a ustar header block. | 
|  | """ | 
|  | info["magic"] = POSIX_MAGIC | 
|  |  | 
|  | if len(info["linkname"].encode(encoding, errors)) > LENGTH_LINK: | 
|  | raise ValueError("linkname is too long") | 
|  |  | 
|  | if len(info["name"].encode(encoding, errors)) > LENGTH_NAME: | 
|  | info["prefix"], info["name"] = self._posix_split_name(info["name"], encoding, errors) | 
|  |  | 
|  | return self._create_header(info, USTAR_FORMAT, encoding, errors) | 
|  |  | 
|  | def create_gnu_header(self, info, encoding, errors): | 
|  | """Return the object as a GNU header block sequence. | 
|  | """ | 
|  | info["magic"] = GNU_MAGIC | 
|  |  | 
|  | buf = b"" | 
|  | if len(info["linkname"].encode(encoding, errors)) > LENGTH_LINK: | 
|  | buf += self._create_gnu_long_header(info["linkname"], GNUTYPE_LONGLINK, encoding, errors) | 
|  |  | 
|  | if len(info["name"].encode(encoding, errors)) > LENGTH_NAME: | 
|  | buf += self._create_gnu_long_header(info["name"], GNUTYPE_LONGNAME, encoding, errors) | 
|  |  | 
|  | return buf + self._create_header(info, GNU_FORMAT, encoding, errors) | 
|  |  | 
|  | def create_pax_header(self, info, encoding): | 
|  | """Return the object as a ustar header block. If it cannot be | 
|  | represented this way, prepend a pax extended header sequence | 
|  | with supplement information. | 
|  | """ | 
|  | info["magic"] = POSIX_MAGIC | 
|  | pax_headers = self.pax_headers.copy() | 
|  |  | 
|  | # Test string fields for values that exceed the field length or cannot | 
|  | # be represented in ASCII encoding. | 
|  | for name, hname, length in ( | 
|  | ("name", "path", LENGTH_NAME), ("linkname", "linkpath", LENGTH_LINK), | 
|  | ("uname", "uname", 32), ("gname", "gname", 32)): | 
|  |  | 
|  | if hname in pax_headers: | 
|  | # The pax header has priority. | 
|  | continue | 
|  |  | 
|  | # Try to encode the string as ASCII. | 
|  | try: | 
|  | info[name].encode("ascii", "strict") | 
|  | except UnicodeEncodeError: | 
|  | pax_headers[hname] = info[name] | 
|  | continue | 
|  |  | 
|  | if len(info[name]) > length: | 
|  | pax_headers[hname] = info[name] | 
|  |  | 
|  | # Test number fields for values that exceed the field limit or values | 
|  | # that like to be stored as float. | 
|  | for name, digits in (("uid", 8), ("gid", 8), ("size", 12), ("mtime", 12)): | 
|  | if name in pax_headers: | 
|  | # The pax header has priority. Avoid overflow. | 
|  | info[name] = 0 | 
|  | continue | 
|  |  | 
|  | val = info[name] | 
|  | if not 0 <= val < 8 ** (digits - 1) or isinstance(val, float): | 
|  | pax_headers[name] = str(val) | 
|  | info[name] = 0 | 
|  |  | 
|  | # Create a pax extended header if necessary. | 
|  | if pax_headers: | 
|  | buf = self._create_pax_generic_header(pax_headers, XHDTYPE, encoding) | 
|  | else: | 
|  | buf = b"" | 
|  |  | 
|  | return buf + self._create_header(info, USTAR_FORMAT, "ascii", "replace") | 
|  |  | 
|  | @classmethod | 
|  | def create_pax_global_header(cls, pax_headers): | 
|  | """Return the object as a pax global header block sequence. | 
|  | """ | 
|  | return cls._create_pax_generic_header(pax_headers, XGLTYPE, "utf-8") | 
|  |  | 
|  | def _posix_split_name(self, name, encoding, errors): | 
|  | """Split a name longer than 100 chars into a prefix | 
|  | and a name part. | 
|  | """ | 
|  | components = name.split("/") | 
|  | for i in range(1, len(components)): | 
|  | prefix = "/".join(components[:i]) | 
|  | name = "/".join(components[i:]) | 
|  | if len(prefix.encode(encoding, errors)) <= LENGTH_PREFIX and \ | 
|  | len(name.encode(encoding, errors)) <= LENGTH_NAME: | 
|  | break | 
|  | else: | 
|  | raise ValueError("name is too long") | 
|  |  | 
|  | return prefix, name | 
|  |  | 
|  | @staticmethod | 
|  | def _create_header(info, format, encoding, errors): | 
|  | """Return a header block. info is a dictionary with file | 
|  | information, format must be one of the *_FORMAT constants. | 
|  | """ | 
|  | parts = [ | 
|  | stn(info.get("name", ""), 100, encoding, errors), | 
|  | itn(info.get("mode", 0) & 0o7777, 8, format), | 
|  | itn(info.get("uid", 0), 8, format), | 
|  | itn(info.get("gid", 0), 8, format), | 
|  | itn(info.get("size", 0), 12, format), | 
|  | itn(info.get("mtime", 0), 12, format), | 
|  | b"        ", # checksum field | 
|  | info.get("type", REGTYPE), | 
|  | stn(info.get("linkname", ""), 100, encoding, errors), | 
|  | info.get("magic", POSIX_MAGIC), | 
|  | stn(info.get("uname", ""), 32, encoding, errors), | 
|  | stn(info.get("gname", ""), 32, encoding, errors), | 
|  | itn(info.get("devmajor", 0), 8, format), | 
|  | itn(info.get("devminor", 0), 8, format), | 
|  | stn(info.get("prefix", ""), 155, encoding, errors) | 
|  | ] | 
|  |  | 
|  | buf = struct.pack("%ds" % BLOCKSIZE, b"".join(parts)) | 
|  | chksum = calc_chksums(buf[-BLOCKSIZE:])[0] | 
|  | buf = buf[:-364] + bytes("%06o\0" % chksum, "ascii") + buf[-357:] | 
|  | return buf | 
|  |  | 
|  | @staticmethod | 
|  | def _create_payload(payload): | 
|  | """Return the string payload filled with zero bytes | 
|  | up to the next 512 byte border. | 
|  | """ | 
|  | blocks, remainder = divmod(len(payload), BLOCKSIZE) | 
|  | if remainder > 0: | 
|  | payload += (BLOCKSIZE - remainder) * NUL | 
|  | return payload | 
|  |  | 
|  | @classmethod | 
|  | def _create_gnu_long_header(cls, name, type, encoding, errors): | 
|  | """Return a GNUTYPE_LONGNAME or GNUTYPE_LONGLINK sequence | 
|  | for name. | 
|  | """ | 
|  | name = name.encode(encoding, errors) + NUL | 
|  |  | 
|  | info = {} | 
|  | info["name"] = "././@LongLink" | 
|  | info["type"] = type | 
|  | info["size"] = len(name) | 
|  | info["magic"] = GNU_MAGIC | 
|  |  | 
|  | # create extended header + name blocks. | 
|  | return cls._create_header(info, USTAR_FORMAT, encoding, errors) + \ | 
|  | cls._create_payload(name) | 
|  |  | 
|  | @classmethod | 
|  | def _create_pax_generic_header(cls, pax_headers, type, encoding): | 
|  | """Return a POSIX.1-2008 extended or global header sequence | 
|  | that contains a list of keyword, value pairs. The values | 
|  | must be strings. | 
|  | """ | 
|  | # Check if one of the fields contains surrogate characters and thereby | 
|  | # forces hdrcharset=BINARY, see _proc_pax() for more information. | 
|  | binary = False | 
|  | for keyword, value in pax_headers.items(): | 
|  | try: | 
|  | value.encode("utf-8", "strict") | 
|  | except UnicodeEncodeError: | 
|  | binary = True | 
|  | break | 
|  |  | 
|  | records = b"" | 
|  | if binary: | 
|  | # Put the hdrcharset field at the beginning of the header. | 
|  | records += b"21 hdrcharset=BINARY\n" | 
|  |  | 
|  | for keyword, value in pax_headers.items(): | 
|  | keyword = keyword.encode("utf-8") | 
|  | if binary: | 
|  | # Try to restore the original byte representation of `value'. | 
|  | # Needless to say, that the encoding must match the string. | 
|  | value = value.encode(encoding, "surrogateescape") | 
|  | else: | 
|  | value = value.encode("utf-8") | 
|  |  | 
|  | l = len(keyword) + len(value) + 3   # ' ' + '=' + '\n' | 
|  | n = p = 0 | 
|  | while True: | 
|  | n = l + len(str(p)) | 
|  | if n == p: | 
|  | break | 
|  | p = n | 
|  | records += bytes(str(p), "ascii") + b" " + keyword + b"=" + value + b"\n" | 
|  |  | 
|  | # We use a hardcoded "././@PaxHeader" name like star does | 
|  | # instead of the one that POSIX recommends. | 
|  | info = {} | 
|  | info["name"] = "././@PaxHeader" | 
|  | info["type"] = type | 
|  | info["size"] = len(records) | 
|  | info["magic"] = POSIX_MAGIC | 
|  |  | 
|  | # Create pax header + record blocks. | 
|  | return cls._create_header(info, USTAR_FORMAT, "ascii", "replace") + \ | 
|  | cls._create_payload(records) | 
|  |  | 
|  | @classmethod | 
|  | def frombuf(cls, buf, encoding, errors): | 
|  | """Construct a TarInfo object from a 512 byte bytes object. | 
|  | """ | 
|  | if len(buf) == 0: | 
|  | raise EmptyHeaderError("empty header") | 
|  | if len(buf) != BLOCKSIZE: | 
|  | raise TruncatedHeaderError("truncated header") | 
|  | if buf.count(NUL) == BLOCKSIZE: | 
|  | raise EOFHeaderError("end of file header") | 
|  |  | 
|  | chksum = nti(buf[148:156]) | 
|  | if chksum not in calc_chksums(buf): | 
|  | raise InvalidHeaderError("bad checksum") | 
|  |  | 
|  | obj = cls() | 
|  | obj.name = nts(buf[0:100], encoding, errors) | 
|  | obj.mode = nti(buf[100:108]) | 
|  | obj.uid = nti(buf[108:116]) | 
|  | obj.gid = nti(buf[116:124]) | 
|  | obj.size = nti(buf[124:136]) | 
|  | obj.mtime = nti(buf[136:148]) | 
|  | obj.chksum = chksum | 
|  | obj.type = buf[156:157] | 
|  | obj.linkname = nts(buf[157:257], encoding, errors) | 
|  | obj.uname = nts(buf[265:297], encoding, errors) | 
|  | obj.gname = nts(buf[297:329], encoding, errors) | 
|  | obj.devmajor = nti(buf[329:337]) | 
|  | obj.devminor = nti(buf[337:345]) | 
|  | prefix = nts(buf[345:500], encoding, errors) | 
|  |  | 
|  | # Old V7 tar format represents a directory as a regular | 
|  | # file with a trailing slash. | 
|  | if obj.type == AREGTYPE and obj.name.endswith("/"): | 
|  | obj.type = DIRTYPE | 
|  |  | 
|  | # The old GNU sparse format occupies some of the unused | 
|  | # space in the buffer for up to 4 sparse structures. | 
|  | # Save them for later processing in _proc_sparse(). | 
|  | if obj.type == GNUTYPE_SPARSE: | 
|  | pos = 386 | 
|  | structs = [] | 
|  | for i in range(4): | 
|  | try: | 
|  | offset = nti(buf[pos:pos + 12]) | 
|  | numbytes = nti(buf[pos + 12:pos + 24]) | 
|  | except ValueError: | 
|  | break | 
|  | structs.append((offset, numbytes)) | 
|  | pos += 24 | 
|  | isextended = bool(buf[482]) | 
|  | origsize = nti(buf[483:495]) | 
|  | obj._sparse_structs = (structs, isextended, origsize) | 
|  |  | 
|  | # Remove redundant slashes from directories. | 
|  | if obj.isdir(): | 
|  | obj.name = obj.name.rstrip("/") | 
|  |  | 
|  | # Reconstruct a ustar longname. | 
|  | if prefix and obj.type not in GNU_TYPES: | 
|  | obj.name = prefix + "/" + obj.name | 
|  | return obj | 
|  |  | 
|  | @classmethod | 
|  | def fromtarfile(cls, tarfile): | 
|  | """Return the next TarInfo object from TarFile object | 
|  | tarfile. | 
|  | """ | 
|  | buf = tarfile.fileobj.read(BLOCKSIZE) | 
|  | obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors) | 
|  | obj.offset = tarfile.fileobj.tell() - BLOCKSIZE | 
|  | return obj._proc_member(tarfile) | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | # The following are methods that are called depending on the type of a | 
|  | # member. The entry point is _proc_member() which can be overridden in a | 
|  | # subclass to add custom _proc_*() methods. A _proc_*() method MUST | 
|  | # implement the following | 
|  | # operations: | 
|  | # 1. Set self.offset_data to the position where the data blocks begin, | 
|  | #    if there is data that follows. | 
|  | # 2. Set tarfile.offset to the position where the next member's header will | 
|  | #    begin. | 
|  | # 3. Return self or another valid TarInfo object. | 
|  | def _proc_member(self, tarfile): | 
|  | """Choose the right processing method depending on | 
|  | the type and call it. | 
|  | """ | 
|  | if self.type in (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK): | 
|  | return self._proc_gnulong(tarfile) | 
|  | elif self.type == GNUTYPE_SPARSE: | 
|  | return self._proc_sparse(tarfile) | 
|  | elif self.type in (XHDTYPE, XGLTYPE, SOLARIS_XHDTYPE): | 
|  | return self._proc_pax(tarfile) | 
|  | else: | 
|  | return self._proc_builtin(tarfile) | 
|  |  | 
|  | def _proc_builtin(self, tarfile): | 
|  | """Process a builtin type or an unknown type which | 
|  | will be treated as a regular file. | 
|  | """ | 
|  | self.offset_data = tarfile.fileobj.tell() | 
|  | offset = self.offset_data | 
|  | if self.isreg() or self.type not in SUPPORTED_TYPES: | 
|  | # Skip the following data blocks. | 
|  | offset += self._block(self.size) | 
|  | tarfile.offset = offset | 
|  |  | 
|  | # Patch the TarInfo object with saved global | 
|  | # header information. | 
|  | self._apply_pax_info(tarfile.pax_headers, tarfile.encoding, tarfile.errors) | 
|  |  | 
|  | return self | 
|  |  | 
|  | def _proc_gnulong(self, tarfile): | 
|  | """Process the blocks that hold a GNU longname | 
|  | or longlink member. | 
|  | """ | 
|  | buf = tarfile.fileobj.read(self._block(self.size)) | 
|  |  | 
|  | # Fetch the next header and process it. | 
|  | try: | 
|  | next = self.fromtarfile(tarfile) | 
|  | except HeaderError: | 
|  | raise SubsequentHeaderError("missing or bad subsequent header") | 
|  |  | 
|  | # Patch the TarInfo object from the next header with | 
|  | # the longname information. | 
|  | next.offset = self.offset | 
|  | if self.type == GNUTYPE_LONGNAME: | 
|  | next.name = nts(buf, tarfile.encoding, tarfile.errors) | 
|  | elif self.type == GNUTYPE_LONGLINK: | 
|  | next.linkname = nts(buf, tarfile.encoding, tarfile.errors) | 
|  |  | 
|  | return next | 
|  |  | 
|  | def _proc_sparse(self, tarfile): | 
|  | """Process a GNU sparse header plus extra headers. | 
|  | """ | 
|  | # We already collected some sparse structures in frombuf(). | 
|  | structs, isextended, origsize = self._sparse_structs | 
|  | del self._sparse_structs | 
|  |  | 
|  | # Collect sparse structures from extended header blocks. | 
|  | while isextended: | 
|  | buf = tarfile.fileobj.read(BLOCKSIZE) | 
|  | pos = 0 | 
|  | for i in range(21): | 
|  | try: | 
|  | offset = nti(buf[pos:pos + 12]) | 
|  | numbytes = nti(buf[pos + 12:pos + 24]) | 
|  | except ValueError: | 
|  | break | 
|  | if offset and numbytes: | 
|  | structs.append((offset, numbytes)) | 
|  | pos += 24 | 
|  | isextended = bool(buf[504]) | 
|  | self.sparse = structs | 
|  |  | 
|  | self.offset_data = tarfile.fileobj.tell() | 
|  | tarfile.offset = self.offset_data + self._block(self.size) | 
|  | self.size = origsize | 
|  | return self | 
|  |  | 
|  | def _proc_pax(self, tarfile): | 
|  | """Process an extended or global header as described in | 
|  | POSIX.1-2008. | 
|  | """ | 
|  | # Read the header information. | 
|  | buf = tarfile.fileobj.read(self._block(self.size)) | 
|  |  | 
|  | # A pax header stores supplemental information for either | 
|  | # the following file (extended) or all following files | 
|  | # (global). | 
|  | if self.type == XGLTYPE: | 
|  | pax_headers = tarfile.pax_headers | 
|  | else: | 
|  | pax_headers = tarfile.pax_headers.copy() | 
|  |  | 
|  | # Check if the pax header contains a hdrcharset field. This tells us | 
|  | # the encoding of the path, linkpath, uname and gname fields. Normally, | 
|  | # these fields are UTF-8 encoded but since POSIX.1-2008 tar | 
|  | # implementations are allowed to store them as raw binary strings if | 
|  | # the translation to UTF-8 fails. | 
|  | match = re.search(br"\d+ hdrcharset=([^\n]+)\n", buf) | 
|  | if match is not None: | 
|  | pax_headers["hdrcharset"] = match.group(1).decode("utf-8") | 
|  |  | 
|  | # For the time being, we don't care about anything other than "BINARY". | 
|  | # The only other value that is currently allowed by the standard is | 
|  | # "ISO-IR 10646 2000 UTF-8" in other words UTF-8. | 
|  | hdrcharset = pax_headers.get("hdrcharset") | 
|  | if hdrcharset == "BINARY": | 
|  | encoding = tarfile.encoding | 
|  | else: | 
|  | encoding = "utf-8" | 
|  |  | 
|  | # Parse pax header information. A record looks like that: | 
|  | # "%d %s=%s\n" % (length, keyword, value). length is the size | 
|  | # of the complete record including the length field itself and | 
|  | # the newline. keyword and value are both UTF-8 encoded strings. | 
|  | regex = re.compile(br"(\d+) ([^=]+)=") | 
|  | pos = 0 | 
|  | while True: | 
|  | match = regex.match(buf, pos) | 
|  | if not match: | 
|  | break | 
|  |  | 
|  | length, keyword = match.groups() | 
|  | length = int(length) | 
|  | if length == 0: | 
|  | raise InvalidHeaderError("invalid header") | 
|  | value = buf[match.end(2) + 1:match.start(1) + length - 1] | 
|  |  | 
|  | # Normally, we could just use "utf-8" as the encoding and "strict" | 
|  | # as the error handler, but we better not take the risk. For | 
|  | # example, GNU tar <= 1.23 is known to store filenames it cannot | 
|  | # translate to UTF-8 as raw strings (unfortunately without a | 
|  | # hdrcharset=BINARY header). | 
|  | # We first try the strict standard encoding, and if that fails we | 
|  | # fall back on the user's encoding and error handler. | 
|  | keyword = self._decode_pax_field(keyword, "utf-8", "utf-8", | 
|  | tarfile.errors) | 
|  | if keyword in PAX_NAME_FIELDS: | 
|  | value = self._decode_pax_field(value, encoding, tarfile.encoding, | 
|  | tarfile.errors) | 
|  | else: | 
|  | value = self._decode_pax_field(value, "utf-8", "utf-8", | 
|  | tarfile.errors) | 
|  |  | 
|  | pax_headers[keyword] = value | 
|  | pos += length | 
|  |  | 
|  | # Fetch the next header. | 
|  | try: | 
|  | next = self.fromtarfile(tarfile) | 
|  | except HeaderError: | 
|  | raise SubsequentHeaderError("missing or bad subsequent header") | 
|  |  | 
|  | # Process GNU sparse information. | 
|  | if "GNU.sparse.map" in pax_headers: | 
|  | # GNU extended sparse format version 0.1. | 
|  | self._proc_gnusparse_01(next, pax_headers) | 
|  |  | 
|  | elif "GNU.sparse.size" in pax_headers: | 
|  | # GNU extended sparse format version 0.0. | 
|  | self._proc_gnusparse_00(next, pax_headers, buf) | 
|  |  | 
|  | elif pax_headers.get("GNU.sparse.major") == "1" and pax_headers.get("GNU.sparse.minor") == "0": | 
|  | # GNU extended sparse format version 1.0. | 
|  | self._proc_gnusparse_10(next, pax_headers, tarfile) | 
|  |  | 
|  | if self.type in (XHDTYPE, SOLARIS_XHDTYPE): | 
|  | # Patch the TarInfo object with the extended header info. | 
|  | next._apply_pax_info(pax_headers, tarfile.encoding, tarfile.errors) | 
|  | next.offset = self.offset | 
|  |  | 
|  | if "size" in pax_headers: | 
|  | # If the extended header replaces the size field, | 
|  | # we need to recalculate the offset where the next | 
|  | # header starts. | 
|  | offset = next.offset_data | 
|  | if next.isreg() or next.type not in SUPPORTED_TYPES: | 
|  | offset += next._block(next.size) | 
|  | tarfile.offset = offset | 
|  |  | 
|  | return next | 
|  |  | 
|  | def _proc_gnusparse_00(self, next, pax_headers, buf): | 
|  | """Process a GNU tar extended sparse header, version 0.0. | 
|  | """ | 
|  | offsets = [] | 
|  | for match in re.finditer(br"\d+ GNU.sparse.offset=(\d+)\n", buf): | 
|  | offsets.append(int(match.group(1))) | 
|  | numbytes = [] | 
|  | for match in re.finditer(br"\d+ GNU.sparse.numbytes=(\d+)\n", buf): | 
|  | numbytes.append(int(match.group(1))) | 
|  | next.sparse = list(zip(offsets, numbytes)) | 
|  |  | 
|  | def _proc_gnusparse_01(self, next, pax_headers): | 
|  | """Process a GNU tar extended sparse header, version 0.1. | 
|  | """ | 
|  | sparse = [int(x) for x in pax_headers["GNU.sparse.map"].split(",")] | 
|  | next.sparse = list(zip(sparse[::2], sparse[1::2])) | 
|  |  | 
|  | def _proc_gnusparse_10(self, next, pax_headers, tarfile): | 
|  | """Process a GNU tar extended sparse header, version 1.0. | 
|  | """ | 
|  | fields = None | 
|  | sparse = [] | 
|  | buf = tarfile.fileobj.read(BLOCKSIZE) | 
|  | fields, buf = buf.split(b"\n", 1) | 
|  | fields = int(fields) | 
|  | while len(sparse) < fields * 2: | 
|  | if b"\n" not in buf: | 
|  | buf += tarfile.fileobj.read(BLOCKSIZE) | 
|  | number, buf = buf.split(b"\n", 1) | 
|  | sparse.append(int(number)) | 
|  | next.offset_data = tarfile.fileobj.tell() | 
|  | next.sparse = list(zip(sparse[::2], sparse[1::2])) | 
|  |  | 
|  | def _apply_pax_info(self, pax_headers, encoding, errors): | 
|  | """Replace fields with supplemental information from a previous | 
|  | pax extended or global header. | 
|  | """ | 
|  | for keyword, value in pax_headers.items(): | 
|  | if keyword == "GNU.sparse.name": | 
|  | setattr(self, "path", value) | 
|  | elif keyword == "GNU.sparse.size": | 
|  | setattr(self, "size", int(value)) | 
|  | elif keyword == "GNU.sparse.realsize": | 
|  | setattr(self, "size", int(value)) | 
|  | elif keyword in PAX_FIELDS: | 
|  | if keyword in PAX_NUMBER_FIELDS: | 
|  | try: | 
|  | value = PAX_NUMBER_FIELDS[keyword](value) | 
|  | except ValueError: | 
|  | value = 0 | 
|  | if keyword == "path": | 
|  | value = value.rstrip("/") | 
|  | setattr(self, keyword, value) | 
|  |  | 
|  | self.pax_headers = pax_headers.copy() | 
|  |  | 
|  | def _decode_pax_field(self, value, encoding, fallback_encoding, fallback_errors): | 
|  | """Decode a single field from a pax record. | 
|  | """ | 
|  | try: | 
|  | return value.decode(encoding, "strict") | 
|  | except UnicodeDecodeError: | 
|  | return value.decode(fallback_encoding, fallback_errors) | 
|  |  | 
|  | def _block(self, count): | 
|  | """Round up a byte count by BLOCKSIZE and return it, | 
|  | e.g. _block(834) => 1024. | 
|  | """ | 
|  | blocks, remainder = divmod(count, BLOCKSIZE) | 
|  | if remainder: | 
|  | blocks += 1 | 
|  | return blocks * BLOCKSIZE | 
|  |  | 
|  | def isreg(self): | 
|  | 'Return True if the Tarinfo object is a regular file.' | 
|  | return self.type in REGULAR_TYPES | 
|  |  | 
|  | def isfile(self): | 
|  | 'Return True if the Tarinfo object is a regular file.' | 
|  | return self.isreg() | 
|  |  | 
|  | def isdir(self): | 
|  | 'Return True if it is a directory.' | 
|  | return self.type == DIRTYPE | 
|  |  | 
|  | def issym(self): | 
|  | 'Return True if it is a symbolic link.' | 
|  | return self.type == SYMTYPE | 
|  |  | 
|  | def islnk(self): | 
|  | 'Return True if it is a hard link.' | 
|  | return self.type == LNKTYPE | 
|  |  | 
|  | def ischr(self): | 
|  | 'Return True if it is a character device.' | 
|  | return self.type == CHRTYPE | 
|  |  | 
|  | def isblk(self): | 
|  | 'Return True if it is a block device.' | 
|  | return self.type == BLKTYPE | 
|  |  | 
|  | def isfifo(self): | 
|  | 'Return True if it is a FIFO.' | 
|  | return self.type == FIFOTYPE | 
|  |  | 
|  | def issparse(self): | 
|  | return self.sparse is not None | 
|  |  | 
|  | def isdev(self): | 
|  | 'Return True if it is one of character device, block device or FIFO.' | 
|  | return self.type in (CHRTYPE, BLKTYPE, FIFOTYPE) | 
|  | # class TarInfo | 
|  |  | 
|  | class TarFile(object): | 
|  | """The TarFile Class provides an interface to tar archives. | 
|  | """ | 
|  |  | 
|  | debug = 0                   # May be set from 0 (no msgs) to 3 (all msgs) | 
|  |  | 
|  | dereference = False         # If true, add content of linked file to the | 
|  | # tar file, else the link. | 
|  |  | 
|  | ignore_zeros = False        # If true, skips empty or invalid blocks and | 
|  | # continues processing. | 
|  |  | 
|  | errorlevel = 1              # If 0, fatal errors only appear in debug | 
|  | # messages (if debug >= 0). If > 0, errors | 
|  | # are passed to the caller as exceptions. | 
|  |  | 
|  | format = DEFAULT_FORMAT     # The format to use when creating an archive. | 
|  |  | 
|  | encoding = ENCODING         # Encoding for 8-bit character strings. | 
|  |  | 
|  | errors = None               # Error handler for unicode conversion. | 
|  |  | 
|  | tarinfo = TarInfo           # The default TarInfo class to use. | 
|  |  | 
|  | fileobject = ExFileObject   # The file-object for extractfile(). | 
|  |  | 
|  | def __init__(self, name=None, mode="r", fileobj=None, format=None, | 
|  | tarinfo=None, dereference=None, ignore_zeros=None, encoding=None, | 
|  | errors="surrogateescape", pax_headers=None, debug=None, | 
|  | errorlevel=None, copybufsize=None): | 
|  | """Open an (uncompressed) tar archive `name'. `mode' is either 'r' to | 
|  | read from an existing archive, 'a' to append data to an existing | 
|  | file or 'w' to create a new file overwriting an existing one. `mode' | 
|  | defaults to 'r'. | 
|  | If `fileobj' is given, it is used for reading or writing data. If it | 
|  | can be determined, `mode' is overridden by `fileobj's mode. | 
|  | `fileobj' is not closed, when TarFile is closed. | 
|  | """ | 
|  | modes = {"r": "rb", "a": "r+b", "w": "wb", "x": "xb"} | 
|  | if mode not in modes: | 
|  | raise ValueError("mode must be 'r', 'a', 'w' or 'x'") | 
|  | self.mode = mode | 
|  | self._mode = modes[mode] | 
|  |  | 
|  | if not fileobj: | 
|  | if self.mode == "a" and not os.path.exists(name): | 
|  | # Create nonexistent files in append mode. | 
|  | self.mode = "w" | 
|  | self._mode = "wb" | 
|  | fileobj = bltn_open(name, self._mode) | 
|  | self._extfileobj = False | 
|  | else: | 
|  | if (name is None and hasattr(fileobj, "name") and | 
|  | isinstance(fileobj.name, (str, bytes))): | 
|  | name = fileobj.name | 
|  | if hasattr(fileobj, "mode"): | 
|  | self._mode = fileobj.mode | 
|  | self._extfileobj = True | 
|  | self.name = os.path.abspath(name) if name else None | 
|  | self.fileobj = fileobj | 
|  |  | 
|  | # Init attributes. | 
|  | if format is not None: | 
|  | self.format = format | 
|  | if tarinfo is not None: | 
|  | self.tarinfo = tarinfo | 
|  | if dereference is not None: | 
|  | self.dereference = dereference | 
|  | if ignore_zeros is not None: | 
|  | self.ignore_zeros = ignore_zeros | 
|  | if encoding is not None: | 
|  | self.encoding = encoding | 
|  | self.errors = errors | 
|  |  | 
|  | if pax_headers is not None and self.format == PAX_FORMAT: | 
|  | self.pax_headers = pax_headers | 
|  | else: | 
|  | self.pax_headers = {} | 
|  |  | 
|  | if debug is not None: | 
|  | self.debug = debug | 
|  | if errorlevel is not None: | 
|  | self.errorlevel = errorlevel | 
|  |  | 
|  | # Init datastructures. | 
|  | self.copybufsize = copybufsize | 
|  | self.closed = False | 
|  | self.members = []       # list of members as TarInfo objects | 
|  | self._loaded = False    # flag if all members have been read | 
|  | self.offset = self.fileobj.tell() | 
|  | # current position in the archive file | 
|  | self.inodes = {}        # dictionary caching the inodes of | 
|  | # archive members already added | 
|  |  | 
|  | try: | 
|  | if self.mode == "r": | 
|  | self.firstmember = None | 
|  | self.firstmember = self.next() | 
|  |  | 
|  | if self.mode == "a": | 
|  | # Move to the end of the archive, | 
|  | # before the first empty block. | 
|  | while True: | 
|  | self.fileobj.seek(self.offset) | 
|  | try: | 
|  | tarinfo = self.tarinfo.fromtarfile(self) | 
|  | self.members.append(tarinfo) | 
|  | except EOFHeaderError: | 
|  | self.fileobj.seek(self.offset) | 
|  | break | 
|  | except HeaderError as e: | 
|  | raise ReadError(str(e)) | 
|  |  | 
|  | if self.mode in ("a", "w", "x"): | 
|  | self._loaded = True | 
|  |  | 
|  | if self.pax_headers: | 
|  | buf = self.tarinfo.create_pax_global_header(self.pax_headers.copy()) | 
|  | self.fileobj.write(buf) | 
|  | self.offset += len(buf) | 
|  | except: | 
|  | if not self._extfileobj: | 
|  | self.fileobj.close() | 
|  | self.closed = True | 
|  | raise | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | # Below are the classmethods which act as alternate constructors to the | 
|  | # TarFile class. The open() method is the only one that is needed for | 
|  | # public use; it is the "super"-constructor and is able to select an | 
|  | # adequate "sub"-constructor for a particular compression using the mapping | 
|  | # from OPEN_METH. | 
|  | # | 
|  | # This concept allows one to subclass TarFile without losing the comfort of | 
|  | # the super-constructor. A sub-constructor is registered and made available | 
|  | # by adding it to the mapping in OPEN_METH. | 
|  |  | 
|  | @classmethod | 
|  | def open(cls, name=None, mode="r", fileobj=None, bufsize=RECORDSIZE, **kwargs): | 
|  | """Open a tar archive for reading, writing or appending. Return | 
|  | an appropriate TarFile class. | 
|  |  | 
|  | mode: | 
|  | 'r' or 'r:*' open for reading with transparent compression | 
|  | 'r:'         open for reading exclusively uncompressed | 
|  | 'r:gz'       open for reading with gzip compression | 
|  | 'r:bz2'      open for reading with bzip2 compression | 
|  | 'r:xz'       open for reading with lzma compression | 
|  | 'a' or 'a:'  open for appending, creating the file if necessary | 
|  | 'w' or 'w:'  open for writing without compression | 
|  | 'w:gz'       open for writing with gzip compression | 
|  | 'w:bz2'      open for writing with bzip2 compression | 
|  | 'w:xz'       open for writing with lzma compression | 
|  |  | 
|  | 'x' or 'x:'  create a tarfile exclusively without compression, raise | 
|  | an exception if the file is already created | 
|  | 'x:gz'       create a gzip compressed tarfile, raise an exception | 
|  | if the file is already created | 
|  | 'x:bz2'      create a bzip2 compressed tarfile, raise an exception | 
|  | if the file is already created | 
|  | 'x:xz'       create an lzma compressed tarfile, raise an exception | 
|  | if the file is already created | 
|  |  | 
|  | 'r|*'        open a stream of tar blocks with transparent compression | 
|  | 'r|'         open an uncompressed stream of tar blocks for reading | 
|  | 'r|gz'       open a gzip compressed stream of tar blocks | 
|  | 'r|bz2'      open a bzip2 compressed stream of tar blocks | 
|  | 'r|xz'       open an lzma compressed stream of tar blocks | 
|  | 'w|'         open an uncompressed stream for writing | 
|  | 'w|gz'       open a gzip compressed stream for writing | 
|  | 'w|bz2'      open a bzip2 compressed stream for writing | 
|  | 'w|xz'       open an lzma compressed stream for writing | 
|  | """ | 
|  |  | 
|  | if not name and not fileobj: | 
|  | raise ValueError("nothing to open") | 
|  |  | 
|  | if mode in ("r", "r:*"): | 
|  | # Find out which *open() is appropriate for opening the file. | 
|  | def not_compressed(comptype): | 
|  | return cls.OPEN_METH[comptype] == 'taropen' | 
|  | for comptype in sorted(cls.OPEN_METH, key=not_compressed): | 
|  | func = getattr(cls, cls.OPEN_METH[comptype]) | 
|  | if fileobj is not None: | 
|  | saved_pos = fileobj.tell() | 
|  | try: | 
|  | return func(name, "r", fileobj, **kwargs) | 
|  | except (ReadError, CompressionError): | 
|  | if fileobj is not None: | 
|  | fileobj.seek(saved_pos) | 
|  | continue | 
|  | raise ReadError("file could not be opened successfully") | 
|  |  | 
|  | elif ":" in mode: | 
|  | filemode, comptype = mode.split(":", 1) | 
|  | filemode = filemode or "r" | 
|  | comptype = comptype or "tar" | 
|  |  | 
|  | # Select the *open() function according to | 
|  | # given compression. | 
|  | if comptype in cls.OPEN_METH: | 
|  | func = getattr(cls, cls.OPEN_METH[comptype]) | 
|  | else: | 
|  | raise CompressionError("unknown compression type %r" % comptype) | 
|  | return func(name, filemode, fileobj, **kwargs) | 
|  |  | 
|  | elif "|" in mode: | 
|  | filemode, comptype = mode.split("|", 1) | 
|  | filemode = filemode or "r" | 
|  | comptype = comptype or "tar" | 
|  |  | 
|  | if filemode not in ("r", "w"): | 
|  | raise ValueError("mode must be 'r' or 'w'") | 
|  |  | 
|  | stream = _Stream(name, filemode, comptype, fileobj, bufsize) | 
|  | try: | 
|  | t = cls(name, filemode, stream, **kwargs) | 
|  | except: | 
|  | stream.close() | 
|  | raise | 
|  | t._extfileobj = False | 
|  | return t | 
|  |  | 
|  | elif mode in ("a", "w", "x"): | 
|  | return cls.taropen(name, mode, fileobj, **kwargs) | 
|  |  | 
|  | raise ValueError("undiscernible mode") | 
|  |  | 
|  | @classmethod | 
|  | def taropen(cls, name, mode="r", fileobj=None, **kwargs): | 
|  | """Open uncompressed tar archive name for reading or writing. | 
|  | """ | 
|  | if mode not in ("r", "a", "w", "x"): | 
|  | raise ValueError("mode must be 'r', 'a', 'w' or 'x'") | 
|  | return cls(name, mode, fileobj, **kwargs) | 
|  |  | 
|  | @classmethod | 
|  | def gzopen(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs): | 
|  | """Open gzip compressed tar archive name for reading or writing. | 
|  | Appending is not allowed. | 
|  | """ | 
|  | if mode not in ("r", "w", "x"): | 
|  | raise ValueError("mode must be 'r', 'w' or 'x'") | 
|  |  | 
|  | try: | 
|  | from gzip import GzipFile | 
|  | except ImportError: | 
|  | raise CompressionError("gzip module is not available") | 
|  |  | 
|  | try: | 
|  | fileobj = GzipFile(name, mode + "b", compresslevel, fileobj) | 
|  | except OSError: | 
|  | if fileobj is not None and mode == 'r': | 
|  | raise ReadError("not a gzip file") | 
|  | raise | 
|  |  | 
|  | try: | 
|  | t = cls.taropen(name, mode, fileobj, **kwargs) | 
|  | except OSError: | 
|  | fileobj.close() | 
|  | if mode == 'r': | 
|  | raise ReadError("not a gzip file") | 
|  | raise | 
|  | except: | 
|  | fileobj.close() | 
|  | raise | 
|  | t._extfileobj = False | 
|  | return t | 
|  |  | 
|  | @classmethod | 
|  | def bz2open(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs): | 
|  | """Open bzip2 compressed tar archive name for reading or writing. | 
|  | Appending is not allowed. | 
|  | """ | 
|  | if mode not in ("r", "w", "x"): | 
|  | raise ValueError("mode must be 'r', 'w' or 'x'") | 
|  |  | 
|  | try: | 
|  | from bz2 import BZ2File | 
|  | except ImportError: | 
|  | raise CompressionError("bz2 module is not available") | 
|  |  | 
|  | fileobj = BZ2File(fileobj or name, mode, compresslevel=compresslevel) | 
|  |  | 
|  | try: | 
|  | t = cls.taropen(name, mode, fileobj, **kwargs) | 
|  | except (OSError, EOFError): | 
|  | fileobj.close() | 
|  | if mode == 'r': | 
|  | raise ReadError("not a bzip2 file") | 
|  | raise | 
|  | except: | 
|  | fileobj.close() | 
|  | raise | 
|  | t._extfileobj = False | 
|  | return t | 
|  |  | 
|  | @classmethod | 
|  | def xzopen(cls, name, mode="r", fileobj=None, preset=None, **kwargs): | 
|  | """Open lzma compressed tar archive name for reading or writing. | 
|  | Appending is not allowed. | 
|  | """ | 
|  | if mode not in ("r", "w", "x"): | 
|  | raise ValueError("mode must be 'r', 'w' or 'x'") | 
|  |  | 
|  | try: | 
|  | from lzma import LZMAFile, LZMAError | 
|  | except ImportError: | 
|  | raise CompressionError("lzma module is not available") | 
|  |  | 
|  | fileobj = LZMAFile(fileobj or name, mode, preset=preset) | 
|  |  | 
|  | try: | 
|  | t = cls.taropen(name, mode, fileobj, **kwargs) | 
|  | except (LZMAError, EOFError): | 
|  | fileobj.close() | 
|  | if mode == 'r': | 
|  | raise ReadError("not an lzma file") | 
|  | raise | 
|  | except: | 
|  | fileobj.close() | 
|  | raise | 
|  | t._extfileobj = False | 
|  | return t | 
|  |  | 
|  | # All *open() methods are registered here. | 
|  | OPEN_METH = { | 
|  | "tar": "taropen",   # uncompressed tar | 
|  | "gz":  "gzopen",    # gzip compressed tar | 
|  | "bz2": "bz2open",   # bzip2 compressed tar | 
|  | "xz":  "xzopen"     # lzma compressed tar | 
|  | } | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | # The public methods which TarFile provides: | 
|  |  | 
|  | def close(self): | 
|  | """Close the TarFile. In write-mode, two finishing zero blocks are | 
|  | appended to the archive. | 
|  | """ | 
|  | if self.closed: | 
|  | return | 
|  |  | 
|  | self.closed = True | 
|  | try: | 
|  | if self.mode in ("a", "w", "x"): | 
|  | self.fileobj.write(NUL * (BLOCKSIZE * 2)) | 
|  | self.offset += (BLOCKSIZE * 2) | 
|  | # fill up the end with zero-blocks | 
|  | # (like option -b20 for tar does) | 
|  | blocks, remainder = divmod(self.offset, RECORDSIZE) | 
|  | if remainder > 0: | 
|  | self.fileobj.write(NUL * (RECORDSIZE - remainder)) | 
|  | finally: | 
|  | if not self._extfileobj: | 
|  | self.fileobj.close() | 
|  |  | 
|  | def getmember(self, name): | 
|  | """Return a TarInfo object for member `name'. If `name' can not be | 
|  | found in the archive, KeyError is raised. If a member occurs more | 
|  | than once in the archive, its last occurrence is assumed to be the | 
|  | most up-to-date version. | 
|  | """ | 
|  | tarinfo = self._getmember(name) | 
|  | if tarinfo is None: | 
|  | raise KeyError("filename %r not found" % name) | 
|  | return tarinfo | 
|  |  | 
|  | def getmembers(self): | 
|  | """Return the members of the archive as a list of TarInfo objects. The | 
|  | list has the same order as the members in the archive. | 
|  | """ | 
|  | self._check() | 
|  | if not self._loaded:    # if we want to obtain a list of | 
|  | self._load()        # all members, we first have to | 
|  | # scan the whole archive. | 
|  | return self.members | 
|  |  | 
|  | def getnames(self): | 
|  | """Return the members of the archive as a list of their names. It has | 
|  | the same order as the list returned by getmembers(). | 
|  | """ | 
|  | return [tarinfo.name for tarinfo in self.getmembers()] | 
|  |  | 
|  | def gettarinfo(self, name=None, arcname=None, fileobj=None): | 
|  | """Create a TarInfo object from the result of os.stat or equivalent | 
|  | on an existing file. The file is either named by `name', or | 
|  | specified as a file object `fileobj' with a file descriptor. If | 
|  | given, `arcname' specifies an alternative name for the file in the | 
|  | archive, otherwise, the name is taken from the 'name' attribute of | 
|  | 'fileobj', or the 'name' argument. The name should be a text | 
|  | string. | 
|  | """ | 
|  | self._check("awx") | 
|  |  | 
|  | # When fileobj is given, replace name by | 
|  | # fileobj's real name. | 
|  | if fileobj is not None: | 
|  | name = fileobj.name | 
|  |  | 
|  | # Building the name of the member in the archive. | 
|  | # Backward slashes are converted to forward slashes, | 
|  | # Absolute paths are turned to relative paths. | 
|  | if arcname is None: | 
|  | arcname = name | 
|  | drv, arcname = os.path.splitdrive(arcname) | 
|  | arcname = arcname.replace(os.sep, "/") | 
|  | arcname = arcname.lstrip("/") | 
|  |  | 
|  | # Now, fill the TarInfo object with | 
|  | # information specific for the file. | 
|  | tarinfo = self.tarinfo() | 
|  | tarinfo.tarfile = self  # Not needed | 
|  |  | 
|  | # Use os.stat or os.lstat, depending on if symlinks shall be resolved. | 
|  | if fileobj is None: | 
|  | if not self.dereference: | 
|  | statres = os.lstat(name) | 
|  | else: | 
|  | statres = os.stat(name) | 
|  | else: | 
|  | statres = os.fstat(fileobj.fileno()) | 
|  | linkname = "" | 
|  |  | 
|  | stmd = statres.st_mode | 
|  | if stat.S_ISREG(stmd): | 
|  | inode = (statres.st_ino, statres.st_dev) | 
|  | if not self.dereference and statres.st_nlink > 1 and \ | 
|  | inode in self.inodes and arcname != self.inodes[inode]: | 
|  | # Is it a hardlink to an already | 
|  | # archived file? | 
|  | type = LNKTYPE | 
|  | linkname = self.inodes[inode] | 
|  | else: | 
|  | # The inode is added only if its valid. | 
|  | # For win32 it is always 0. | 
|  | type = REGTYPE | 
|  | if inode[0]: | 
|  | self.inodes[inode] = arcname | 
|  | elif stat.S_ISDIR(stmd): | 
|  | type = DIRTYPE | 
|  | elif stat.S_ISFIFO(stmd): | 
|  | type = FIFOTYPE | 
|  | elif stat.S_ISLNK(stmd): | 
|  | type = SYMTYPE | 
|  | linkname = os.readlink(name) | 
|  | elif stat.S_ISCHR(stmd): | 
|  | type = CHRTYPE | 
|  | elif stat.S_ISBLK(stmd): | 
|  | type = BLKTYPE | 
|  | else: | 
|  | return None | 
|  |  | 
|  | # Fill the TarInfo object with all | 
|  | # information we can get. | 
|  | tarinfo.name = arcname | 
|  | tarinfo.mode = stmd | 
|  | tarinfo.uid = statres.st_uid | 
|  | tarinfo.gid = statres.st_gid | 
|  | if type == REGTYPE: | 
|  | tarinfo.size = statres.st_size | 
|  | else: | 
|  | tarinfo.size = 0 | 
|  | tarinfo.mtime = statres.st_mtime | 
|  | tarinfo.type = type | 
|  | tarinfo.linkname = linkname | 
|  | if pwd: | 
|  | try: | 
|  | tarinfo.uname = pwd.getpwuid(tarinfo.uid)[0] | 
|  | except KeyError: | 
|  | pass | 
|  | if grp: | 
|  | try: | 
|  | tarinfo.gname = grp.getgrgid(tarinfo.gid)[0] | 
|  | except KeyError: | 
|  | pass | 
|  |  | 
|  | if type in (CHRTYPE, BLKTYPE): | 
|  | if hasattr(os, "major") and hasattr(os, "minor"): | 
|  | tarinfo.devmajor = os.major(statres.st_rdev) | 
|  | tarinfo.devminor = os.minor(statres.st_rdev) | 
|  | return tarinfo | 
|  |  | 
|  | def list(self, verbose=True, *, members=None): | 
|  | """Print a table of contents to sys.stdout. If `verbose' is False, only | 
|  | the names of the members are printed. If it is True, an `ls -l'-like | 
|  | output is produced. `members' is optional and must be a subset of the | 
|  | list returned by getmembers(). | 
|  | """ | 
|  | self._check() | 
|  |  | 
|  | if members is None: | 
|  | members = self | 
|  | for tarinfo in members: | 
|  | if verbose: | 
|  | _safe_print(stat.filemode(tarinfo.mode)) | 
|  | _safe_print("%s/%s" % (tarinfo.uname or tarinfo.uid, | 
|  | tarinfo.gname or tarinfo.gid)) | 
|  | if tarinfo.ischr() or tarinfo.isblk(): | 
|  | _safe_print("%10s" % | 
|  | ("%d,%d" % (tarinfo.devmajor, tarinfo.devminor))) | 
|  | else: | 
|  | _safe_print("%10d" % tarinfo.size) | 
|  | _safe_print("%d-%02d-%02d %02d:%02d:%02d" \ | 
|  | % time.localtime(tarinfo.mtime)[:6]) | 
|  |  | 
|  | _safe_print(tarinfo.name + ("/" if tarinfo.isdir() else "")) | 
|  |  | 
|  | if verbose: | 
|  | if tarinfo.issym(): | 
|  | _safe_print("-> " + tarinfo.linkname) | 
|  | if tarinfo.islnk(): | 
|  | _safe_print("link to " + tarinfo.linkname) | 
|  | print() | 
|  |  | 
|  | def add(self, name, arcname=None, recursive=True, *, filter=None): | 
|  | """Add the file `name' to the archive. `name' may be any type of file | 
|  | (directory, fifo, symbolic link, etc.). If given, `arcname' | 
|  | specifies an alternative name for the file in the archive. | 
|  | Directories are added recursively by default. This can be avoided by | 
|  | setting `recursive' to False. `filter' is a function | 
|  | that expects a TarInfo object argument and returns the changed | 
|  | TarInfo object, if it returns None the TarInfo object will be | 
|  | excluded from the archive. | 
|  | """ | 
|  | self._check("awx") | 
|  |  | 
|  | if arcname is None: | 
|  | arcname = name | 
|  |  | 
|  | # Skip if somebody tries to archive the archive... | 
|  | if self.name is not None and os.path.abspath(name) == self.name: | 
|  | self._dbg(2, "tarfile: Skipped %r" % name) | 
|  | return | 
|  |  | 
|  | self._dbg(1, name) | 
|  |  | 
|  | # Create a TarInfo object from the file. | 
|  | tarinfo = self.gettarinfo(name, arcname) | 
|  |  | 
|  | if tarinfo is None: | 
|  | self._dbg(1, "tarfile: Unsupported type %r" % name) | 
|  | return | 
|  |  | 
|  | # Change or exclude the TarInfo object. | 
|  | if filter is not None: | 
|  | tarinfo = filter(tarinfo) | 
|  | if tarinfo is None: | 
|  | self._dbg(2, "tarfile: Excluded %r" % name) | 
|  | return | 
|  |  | 
|  | # Append the tar header and data to the archive. | 
|  | if tarinfo.isreg(): | 
|  | with bltn_open(name, "rb") as f: | 
|  | self.addfile(tarinfo, f) | 
|  |  | 
|  | elif tarinfo.isdir(): | 
|  | self.addfile(tarinfo) | 
|  | if recursive: | 
|  | for f in sorted(os.listdir(name)): | 
|  | self.add(os.path.join(name, f), os.path.join(arcname, f), | 
|  | recursive, filter=filter) | 
|  |  | 
|  | else: | 
|  | self.addfile(tarinfo) | 
|  |  | 
|  | def addfile(self, tarinfo, fileobj=None): | 
|  | """Add the TarInfo object `tarinfo' to the archive. If `fileobj' is | 
|  | given, it should be a binary file, and tarinfo.size bytes are read | 
|  | from it and added to the archive. You can create TarInfo objects | 
|  | directly, or by using gettarinfo(). | 
|  | """ | 
|  | self._check("awx") | 
|  |  | 
|  | tarinfo = copy.copy(tarinfo) | 
|  |  | 
|  | buf = tarinfo.tobuf(self.format, self.encoding, self.errors) | 
|  | self.fileobj.write(buf) | 
|  | self.offset += len(buf) | 
|  | bufsize=self.copybufsize | 
|  | # If there's data to follow, append it. | 
|  | if fileobj is not None: | 
|  | copyfileobj(fileobj, self.fileobj, tarinfo.size, bufsize=bufsize) | 
|  | blocks, remainder = divmod(tarinfo.size, BLOCKSIZE) | 
|  | if remainder > 0: | 
|  | self.fileobj.write(NUL * (BLOCKSIZE - remainder)) | 
|  | blocks += 1 | 
|  | self.offset += blocks * BLOCKSIZE | 
|  |  | 
|  | self.members.append(tarinfo) | 
|  |  | 
|  | def extractall(self, path=".", members=None, *, numeric_owner=False): | 
|  | """Extract all members from the archive to the current working | 
|  | directory and set owner, modification time and permissions on | 
|  | directories afterwards. `path' specifies a different directory | 
|  | to extract to. `members' is optional and must be a subset of the | 
|  | list returned by getmembers(). If `numeric_owner` is True, only | 
|  | the numbers for user/group names are used and not the names. | 
|  | """ | 
|  | directories = [] | 
|  |  | 
|  | if members is None: | 
|  | members = self | 
|  |  | 
|  | for tarinfo in members: | 
|  | if tarinfo.isdir(): | 
|  | # Extract directories with a safe mode. | 
|  | directories.append(tarinfo) | 
|  | tarinfo = copy.copy(tarinfo) | 
|  | tarinfo.mode = 0o700 | 
|  | # Do not set_attrs directories, as we will do that further down | 
|  | self.extract(tarinfo, path, set_attrs=not tarinfo.isdir(), | 
|  | numeric_owner=numeric_owner) | 
|  |  | 
|  | # Reverse sort directories. | 
|  | directories.sort(key=lambda a: a.name) | 
|  | directories.reverse() | 
|  |  | 
|  | # Set correct owner, mtime and filemode on directories. | 
|  | for tarinfo in directories: | 
|  | dirpath = os.path.join(path, tarinfo.name) | 
|  | try: | 
|  | self.chown(tarinfo, dirpath, numeric_owner=numeric_owner) | 
|  | self.utime(tarinfo, dirpath) | 
|  | self.chmod(tarinfo, dirpath) | 
|  | except ExtractError as e: | 
|  | if self.errorlevel > 1: | 
|  | raise | 
|  | else: | 
|  | self._dbg(1, "tarfile: %s" % e) | 
|  |  | 
|  | def extract(self, member, path="", set_attrs=True, *, numeric_owner=False): | 
|  | """Extract a member from the archive to the current working directory, | 
|  | using its full name. Its file information is extracted as accurately | 
|  | as possible. `member' may be a filename or a TarInfo object. You can | 
|  | specify a different directory using `path'. File attributes (owner, | 
|  | mtime, mode) are set unless `set_attrs' is False. If `numeric_owner` | 
|  | is True, only the numbers for user/group names are used and not | 
|  | the names. | 
|  | """ | 
|  | self._check("r") | 
|  |  | 
|  | if isinstance(member, str): | 
|  | tarinfo = self.getmember(member) | 
|  | else: | 
|  | tarinfo = member | 
|  |  | 
|  | # Prepare the link target for makelink(). | 
|  | if tarinfo.islnk(): | 
|  | tarinfo._link_target = os.path.join(path, tarinfo.linkname) | 
|  |  | 
|  | try: | 
|  | self._extract_member(tarinfo, os.path.join(path, tarinfo.name), | 
|  | set_attrs=set_attrs, | 
|  | numeric_owner=numeric_owner) | 
|  | except OSError as e: | 
|  | if self.errorlevel > 0: | 
|  | raise | 
|  | else: | 
|  | if e.filename is None: | 
|  | self._dbg(1, "tarfile: %s" % e.strerror) | 
|  | else: | 
|  | self._dbg(1, "tarfile: %s %r" % (e.strerror, e.filename)) | 
|  | except ExtractError as e: | 
|  | if self.errorlevel > 1: | 
|  | raise | 
|  | else: | 
|  | self._dbg(1, "tarfile: %s" % e) | 
|  |  | 
|  | def extractfile(self, member): | 
|  | """Extract a member from the archive as a file object. `member' may be | 
|  | a filename or a TarInfo object. If `member' is a regular file or a | 
|  | link, an io.BufferedReader object is returned. Otherwise, None is | 
|  | returned. | 
|  | """ | 
|  | self._check("r") | 
|  |  | 
|  | if isinstance(member, str): | 
|  | tarinfo = self.getmember(member) | 
|  | else: | 
|  | tarinfo = member | 
|  |  | 
|  | if tarinfo.isreg() or tarinfo.type not in SUPPORTED_TYPES: | 
|  | # Members with unknown types are treated as regular files. | 
|  | return self.fileobject(self, tarinfo) | 
|  |  | 
|  | elif tarinfo.islnk() or tarinfo.issym(): | 
|  | if isinstance(self.fileobj, _Stream): | 
|  | # A small but ugly workaround for the case that someone tries | 
|  | # to extract a (sym)link as a file-object from a non-seekable | 
|  | # stream of tar blocks. | 
|  | raise StreamError("cannot extract (sym)link as file object") | 
|  | else: | 
|  | # A (sym)link's file object is its target's file object. | 
|  | return self.extractfile(self._find_link_target(tarinfo)) | 
|  | else: | 
|  | # If there's no data associated with the member (directory, chrdev, | 
|  | # blkdev, etc.), return None instead of a file object. | 
|  | return None | 
|  |  | 
|  | def _extract_member(self, tarinfo, targetpath, set_attrs=True, | 
|  | numeric_owner=False): | 
|  | """Extract the TarInfo object tarinfo to a physical | 
|  | file called targetpath. | 
|  | """ | 
|  | # Fetch the TarInfo object for the given name | 
|  | # and build the destination pathname, replacing | 
|  | # forward slashes to platform specific separators. | 
|  | targetpath = targetpath.rstrip("/") | 
|  | targetpath = targetpath.replace("/", os.sep) | 
|  |  | 
|  | # Create all upper directories. | 
|  | upperdirs = os.path.dirname(targetpath) | 
|  | if upperdirs and not os.path.exists(upperdirs): | 
|  | # Create directories that are not part of the archive with | 
|  | # default permissions. | 
|  | os.makedirs(upperdirs) | 
|  |  | 
|  | if tarinfo.islnk() or tarinfo.issym(): | 
|  | self._dbg(1, "%s -> %s" % (tarinfo.name, tarinfo.linkname)) | 
|  | else: | 
|  | self._dbg(1, tarinfo.name) | 
|  |  | 
|  | if tarinfo.isreg(): | 
|  | self.makefile(tarinfo, targetpath) | 
|  | elif tarinfo.isdir(): | 
|  | self.makedir(tarinfo, targetpath) | 
|  | elif tarinfo.isfifo(): | 
|  | self.makefifo(tarinfo, targetpath) | 
|  | elif tarinfo.ischr() or tarinfo.isblk(): | 
|  | self.makedev(tarinfo, targetpath) | 
|  | elif tarinfo.islnk() or tarinfo.issym(): | 
|  | self.makelink(tarinfo, targetpath) | 
|  | elif tarinfo.type not in SUPPORTED_TYPES: | 
|  | self.makeunknown(tarinfo, targetpath) | 
|  | else: | 
|  | self.makefile(tarinfo, targetpath) | 
|  |  | 
|  | if set_attrs: | 
|  | self.chown(tarinfo, targetpath, numeric_owner) | 
|  | if not tarinfo.issym(): | 
|  | self.chmod(tarinfo, targetpath) | 
|  | self.utime(tarinfo, targetpath) | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | # Below are the different file methods. They are called via | 
|  | # _extract_member() when extract() is called. They can be replaced in a | 
|  | # subclass to implement other functionality. | 
|  |  | 
|  | def makedir(self, tarinfo, targetpath): | 
|  | """Make a directory called targetpath. | 
|  | """ | 
|  | try: | 
|  | # Use a safe mode for the directory, the real mode is set | 
|  | # later in _extract_member(). | 
|  | os.mkdir(targetpath, 0o700) | 
|  | except FileExistsError: | 
|  | pass | 
|  |  | 
|  | def makefile(self, tarinfo, targetpath): | 
|  | """Make a file called targetpath. | 
|  | """ | 
|  | source = self.fileobj | 
|  | source.seek(tarinfo.offset_data) | 
|  | bufsize = self.copybufsize | 
|  | with bltn_open(targetpath, "wb") as target: | 
|  | if tarinfo.sparse is not None: | 
|  | for offset, size in tarinfo.sparse: | 
|  | target.seek(offset) | 
|  | copyfileobj(source, target, size, ReadError, bufsize) | 
|  | target.seek(tarinfo.size) | 
|  | target.truncate() | 
|  | else: | 
|  | copyfileobj(source, target, tarinfo.size, ReadError, bufsize) | 
|  |  | 
|  | def makeunknown(self, tarinfo, targetpath): | 
|  | """Make a file from a TarInfo object with an unknown type | 
|  | at targetpath. | 
|  | """ | 
|  | self.makefile(tarinfo, targetpath) | 
|  | self._dbg(1, "tarfile: Unknown file type %r, " \ | 
|  | "extracted as regular file." % tarinfo.type) | 
|  |  | 
|  | def makefifo(self, tarinfo, targetpath): | 
|  | """Make a fifo called targetpath. | 
|  | """ | 
|  | if hasattr(os, "mkfifo"): | 
|  | os.mkfifo(targetpath) | 
|  | else: | 
|  | raise ExtractError("fifo not supported by system") | 
|  |  | 
|  | def makedev(self, tarinfo, targetpath): | 
|  | """Make a character or block device called targetpath. | 
|  | """ | 
|  | if not hasattr(os, "mknod") or not hasattr(os, "makedev"): | 
|  | raise ExtractError("special devices not supported by system") | 
|  |  | 
|  | mode = tarinfo.mode | 
|  | if tarinfo.isblk(): | 
|  | mode |= stat.S_IFBLK | 
|  | else: | 
|  | mode |= stat.S_IFCHR | 
|  |  | 
|  | os.mknod(targetpath, mode, | 
|  | os.makedev(tarinfo.devmajor, tarinfo.devminor)) | 
|  |  | 
|  | def makelink(self, tarinfo, targetpath): | 
|  | """Make a (symbolic) link called targetpath. If it cannot be created | 
|  | (platform limitation), we try to make a copy of the referenced file | 
|  | instead of a link. | 
|  | """ | 
|  | try: | 
|  | # For systems that support symbolic and hard links. | 
|  | if tarinfo.issym(): | 
|  | if os.path.lexists(targetpath): | 
|  | # Avoid FileExistsError on following os.symlink. | 
|  | os.unlink(targetpath) | 
|  | os.symlink(tarinfo.linkname, targetpath) | 
|  | else: | 
|  | # See extract(). | 
|  | if os.path.exists(tarinfo._link_target): | 
|  | os.link(tarinfo._link_target, targetpath) | 
|  | else: | 
|  | self._extract_member(self._find_link_target(tarinfo), | 
|  | targetpath) | 
|  | except symlink_exception: | 
|  | try: | 
|  | self._extract_member(self._find_link_target(tarinfo), | 
|  | targetpath) | 
|  | except KeyError: | 
|  | raise ExtractError("unable to resolve link inside archive") | 
|  |  | 
|  | def chown(self, tarinfo, targetpath, numeric_owner): | 
|  | """Set owner of targetpath according to tarinfo. If numeric_owner | 
|  | is True, use .gid/.uid instead of .gname/.uname. If numeric_owner | 
|  | is False, fall back to .gid/.uid when the search based on name | 
|  | fails. | 
|  | """ | 
|  | if hasattr(os, "geteuid") and os.geteuid() == 0: | 
|  | # We have to be root to do so. | 
|  | g = tarinfo.gid | 
|  | u = tarinfo.uid | 
|  | if not numeric_owner: | 
|  | try: | 
|  | if grp: | 
|  | g = grp.getgrnam(tarinfo.gname)[2] | 
|  | except KeyError: | 
|  | pass | 
|  | try: | 
|  | if pwd: | 
|  | u = pwd.getpwnam(tarinfo.uname)[2] | 
|  | except KeyError: | 
|  | pass | 
|  | try: | 
|  | if tarinfo.issym() and hasattr(os, "lchown"): | 
|  | os.lchown(targetpath, u, g) | 
|  | else: | 
|  | os.chown(targetpath, u, g) | 
|  | except OSError: | 
|  | raise ExtractError("could not change owner") | 
|  |  | 
|  | def chmod(self, tarinfo, targetpath): | 
|  | """Set file permissions of targetpath according to tarinfo. | 
|  | """ | 
|  | try: | 
|  | os.chmod(targetpath, tarinfo.mode) | 
|  | except OSError: | 
|  | raise ExtractError("could not change mode") | 
|  |  | 
|  | def utime(self, tarinfo, targetpath): | 
|  | """Set modification time of targetpath according to tarinfo. | 
|  | """ | 
|  | if not hasattr(os, 'utime'): | 
|  | return | 
|  | try: | 
|  | os.utime(targetpath, (tarinfo.mtime, tarinfo.mtime)) | 
|  | except OSError: | 
|  | raise ExtractError("could not change modification time") | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | def next(self): | 
|  | """Return the next member of the archive as a TarInfo object, when | 
|  | TarFile is opened for reading. Return None if there is no more | 
|  | available. | 
|  | """ | 
|  | self._check("ra") | 
|  | if self.firstmember is not None: | 
|  | m = self.firstmember | 
|  | self.firstmember = None | 
|  | return m | 
|  |  | 
|  | # Advance the file pointer. | 
|  | if self.offset != self.fileobj.tell(): | 
|  | self.fileobj.seek(self.offset - 1) | 
|  | if not self.fileobj.read(1): | 
|  | raise ReadError("unexpected end of data") | 
|  |  | 
|  | # Read the next block. | 
|  | tarinfo = None | 
|  | while True: | 
|  | try: | 
|  | tarinfo = self.tarinfo.fromtarfile(self) | 
|  | except EOFHeaderError as e: | 
|  | if self.ignore_zeros: | 
|  | self._dbg(2, "0x%X: %s" % (self.offset, e)) | 
|  | self.offset += BLOCKSIZE | 
|  | continue | 
|  | except InvalidHeaderError as e: | 
|  | if self.ignore_zeros: | 
|  | self._dbg(2, "0x%X: %s" % (self.offset, e)) | 
|  | self.offset += BLOCKSIZE | 
|  | continue | 
|  | elif self.offset == 0: | 
|  | raise ReadError(str(e)) | 
|  | except EmptyHeaderError: | 
|  | if self.offset == 0: | 
|  | raise ReadError("empty file") | 
|  | except TruncatedHeaderError as e: | 
|  | if self.offset == 0: | 
|  | raise ReadError(str(e)) | 
|  | except SubsequentHeaderError as e: | 
|  | raise ReadError(str(e)) | 
|  | break | 
|  |  | 
|  | if tarinfo is not None: | 
|  | self.members.append(tarinfo) | 
|  | else: | 
|  | self._loaded = True | 
|  |  | 
|  | return tarinfo | 
|  |  | 
|  | #-------------------------------------------------------------------------- | 
|  | # Little helper methods: | 
|  |  | 
|  | def _getmember(self, name, tarinfo=None, normalize=False): | 
|  | """Find an archive member by name from bottom to top. | 
|  | If tarinfo is given, it is used as the starting point. | 
|  | """ | 
|  | # Ensure that all members have been loaded. | 
|  | members = self.getmembers() | 
|  |  | 
|  | # Limit the member search list up to tarinfo. | 
|  | if tarinfo is not None: | 
|  | members = members[:members.index(tarinfo)] | 
|  |  | 
|  | if normalize: | 
|  | name = os.path.normpath(name) | 
|  |  | 
|  | for member in reversed(members): | 
|  | if normalize: | 
|  | member_name = os.path.normpath(member.name) | 
|  | else: | 
|  | member_name = member.name | 
|  |  | 
|  | if name == member_name: | 
|  | return member | 
|  |  | 
|  | def _load(self): | 
|  | """Read through the entire archive file and look for readable | 
|  | members. | 
|  | """ | 
|  | while True: | 
|  | tarinfo = self.next() | 
|  | if tarinfo is None: | 
|  | break | 
|  | self._loaded = True | 
|  |  | 
|  | def _check(self, mode=None): | 
|  | """Check if TarFile is still open, and if the operation's mode | 
|  | corresponds to TarFile's mode. | 
|  | """ | 
|  | if self.closed: | 
|  | raise OSError("%s is closed" % self.__class__.__name__) | 
|  | if mode is not None and self.mode not in mode: | 
|  | raise OSError("bad operation for mode %r" % self.mode) | 
|  |  | 
|  | def _find_link_target(self, tarinfo): | 
|  | """Find the target member of a symlink or hardlink member in the | 
|  | archive. | 
|  | """ | 
|  | if tarinfo.issym(): | 
|  | # Always search the entire archive. | 
|  | linkname = "/".join(filter(None, (os.path.dirname(tarinfo.name), tarinfo.linkname))) | 
|  | limit = None | 
|  | else: | 
|  | # Search the archive before the link, because a hard link is | 
|  | # just a reference to an already archived file. | 
|  | linkname = tarinfo.linkname | 
|  | limit = tarinfo | 
|  |  | 
|  | member = self._getmember(linkname, tarinfo=limit, normalize=True) | 
|  | if member is None: | 
|  | raise KeyError("linkname %r not found" % linkname) | 
|  | return member | 
|  |  | 
|  | def __iter__(self): | 
|  | """Provide an iterator object. | 
|  | """ | 
|  | if self._loaded: | 
|  | yield from self.members | 
|  | return | 
|  |  | 
|  | # Yield items using TarFile's next() method. | 
|  | # When all members have been read, set TarFile as _loaded. | 
|  | index = 0 | 
|  | # Fix for SF #1100429: Under rare circumstances it can | 
|  | # happen that getmembers() is called during iteration, | 
|  | # which will have already exhausted the next() method. | 
|  | if self.firstmember is not None: | 
|  | tarinfo = self.next() | 
|  | index += 1 | 
|  | yield tarinfo | 
|  |  | 
|  | while True: | 
|  | if index < len(self.members): | 
|  | tarinfo = self.members[index] | 
|  | elif not self._loaded: | 
|  | tarinfo = self.next() | 
|  | if not tarinfo: | 
|  | self._loaded = True | 
|  | return | 
|  | else: | 
|  | return | 
|  | index += 1 | 
|  | yield tarinfo | 
|  |  | 
|  | def _dbg(self, level, msg): | 
|  | """Write debugging output to sys.stderr. | 
|  | """ | 
|  | if level <= self.debug: | 
|  | print(msg, file=sys.stderr) | 
|  |  | 
|  | def __enter__(self): | 
|  | self._check() | 
|  | return self | 
|  |  | 
|  | def __exit__(self, type, value, traceback): | 
|  | if type is None: | 
|  | self.close() | 
|  | else: | 
|  | # An exception occurred. We must not call close() because | 
|  | # it would try to write end-of-archive blocks and padding. | 
|  | if not self._extfileobj: | 
|  | self.fileobj.close() | 
|  | self.closed = True | 
|  |  | 
|  | #-------------------- | 
|  | # exported functions | 
|  | #-------------------- | 
|  | def is_tarfile(name): | 
|  | """Return True if name points to a tar archive that we | 
|  | are able to handle, else return False. | 
|  | """ | 
|  | try: | 
|  | t = open(name) | 
|  | t.close() | 
|  | return True | 
|  | except TarError: | 
|  | return False | 
|  |  | 
|  | open = TarFile.open | 
|  |  | 
|  |  | 
|  | def main(): | 
|  | import argparse | 
|  |  | 
|  | description = 'A simple command-line interface for tarfile module.' | 
|  | parser = argparse.ArgumentParser(description=description) | 
|  | parser.add_argument('-v', '--verbose', action='store_true', default=False, | 
|  | help='Verbose output') | 
|  | group = parser.add_mutually_exclusive_group(required=True) | 
|  | group.add_argument('-l', '--list', metavar='<tarfile>', | 
|  | help='Show listing of a tarfile') | 
|  | group.add_argument('-e', '--extract', nargs='+', | 
|  | metavar=('<tarfile>', '<output_dir>'), | 
|  | help='Extract tarfile into target dir') | 
|  | group.add_argument('-c', '--create', nargs='+', | 
|  | metavar=('<name>', '<file>'), | 
|  | help='Create tarfile from sources') | 
|  | group.add_argument('-t', '--test', metavar='<tarfile>', | 
|  | help='Test if a tarfile is valid') | 
|  | args = parser.parse_args() | 
|  |  | 
|  | if args.test is not None: | 
|  | src = args.test | 
|  | if is_tarfile(src): | 
|  | with open(src, 'r') as tar: | 
|  | tar.getmembers() | 
|  | print(tar.getmembers(), file=sys.stderr) | 
|  | if args.verbose: | 
|  | print('{!r} is a tar archive.'.format(src)) | 
|  | else: | 
|  | parser.exit(1, '{!r} is not a tar archive.\n'.format(src)) | 
|  |  | 
|  | elif args.list is not None: | 
|  | src = args.list | 
|  | if is_tarfile(src): | 
|  | with TarFile.open(src, 'r:*') as tf: | 
|  | tf.list(verbose=args.verbose) | 
|  | else: | 
|  | parser.exit(1, '{!r} is not a tar archive.\n'.format(src)) | 
|  |  | 
|  | elif args.extract is not None: | 
|  | if len(args.extract) == 1: | 
|  | src = args.extract[0] | 
|  | curdir = os.curdir | 
|  | elif len(args.extract) == 2: | 
|  | src, curdir = args.extract | 
|  | else: | 
|  | parser.exit(1, parser.format_help()) | 
|  |  | 
|  | if is_tarfile(src): | 
|  | with TarFile.open(src, 'r:*') as tf: | 
|  | tf.extractall(path=curdir) | 
|  | if args.verbose: | 
|  | if curdir == '.': | 
|  | msg = '{!r} file is extracted.'.format(src) | 
|  | else: | 
|  | msg = ('{!r} file is extracted ' | 
|  | 'into {!r} directory.').format(src, curdir) | 
|  | print(msg) | 
|  | else: | 
|  | parser.exit(1, '{!r} is not a tar archive.\n'.format(src)) | 
|  |  | 
|  | elif args.create is not None: | 
|  | tar_name = args.create.pop(0) | 
|  | _, ext = os.path.splitext(tar_name) | 
|  | compressions = { | 
|  | # gz | 
|  | '.gz': 'gz', | 
|  | '.tgz': 'gz', | 
|  | # xz | 
|  | '.xz': 'xz', | 
|  | '.txz': 'xz', | 
|  | # bz2 | 
|  | '.bz2': 'bz2', | 
|  | '.tbz': 'bz2', | 
|  | '.tbz2': 'bz2', | 
|  | '.tb2': 'bz2', | 
|  | } | 
|  | tar_mode = 'w:' + compressions[ext] if ext in compressions else 'w' | 
|  | tar_files = args.create | 
|  |  | 
|  | with TarFile.open(tar_name, tar_mode) as tf: | 
|  | for file_name in tar_files: | 
|  | tf.add(file_name) | 
|  |  | 
|  | if args.verbose: | 
|  | print('{!r} file created.'.format(tar_name)) | 
|  |  | 
|  | if __name__ == '__main__': | 
|  | main() |