|
@@ -28,36 +28,43 @@ from cStringIO import (
|
|
|
import mmap
|
|
|
import os
|
|
|
import stat
|
|
|
-import time
|
|
|
import zlib
|
|
|
|
|
|
from dulwich.errors import (
|
|
|
+ ChecksumMismatch,
|
|
|
NotBlobError,
|
|
|
NotCommitError,
|
|
|
+ NotTagError,
|
|
|
NotTreeError,
|
|
|
+ ObjectFormatException,
|
|
|
)
|
|
|
from dulwich.file import GitFile
|
|
|
from dulwich.misc import (
|
|
|
make_sha,
|
|
|
)
|
|
|
|
|
|
-BLOB_ID = "blob"
|
|
|
-TAG_ID = "tag"
|
|
|
-TREE_ID = "tree"
|
|
|
-COMMIT_ID = "commit"
|
|
|
-PARENT_ID = "parent"
|
|
|
-AUTHOR_ID = "author"
|
|
|
-COMMITTER_ID = "committer"
|
|
|
-OBJECT_ID = "object"
|
|
|
-TYPE_ID = "type"
|
|
|
-TAGGER_ID = "tagger"
|
|
|
-ENCODING_ID = "encoding"
|
|
|
+
|
|
|
+# Header fields for commits
|
|
|
+_TREE_HEADER = "tree"
|
|
|
+_PARENT_HEADER = "parent"
|
|
|
+_AUTHOR_HEADER = "author"
|
|
|
+_COMMITTER_HEADER = "committer"
|
|
|
+_ENCODING_HEADER = "encoding"
|
|
|
+
|
|
|
+
|
|
|
+# Header fields for objects
|
|
|
+_OBJECT_HEADER = "object"
|
|
|
+_TYPE_HEADER = "type"
|
|
|
+_TAG_HEADER = "tag"
|
|
|
+_TAGGER_HEADER = "tagger"
|
|
|
+
|
|
|
|
|
|
S_IFGITLINK = 0160000
|
|
|
|
|
|
def S_ISGITLINK(m):
|
|
|
return (stat.S_IFMT(m) == S_IFGITLINK)
|
|
|
|
|
|
+
|
|
|
def _decompress(string):
|
|
|
dcomp = zlib.decompressobj()
|
|
|
dcomped = dcomp.decompress(string)
|
|
@@ -78,6 +85,27 @@ def hex_to_sha(hex):
|
|
|
return binascii.unhexlify(hex)
|
|
|
|
|
|
|
|
|
+def hex_to_filename(path, hex):
|
|
|
+ """Takes a hex sha and returns its filename relative to the given path."""
|
|
|
+ dir = hex[:2]
|
|
|
+ file = hex[2:]
|
|
|
+ # Check from object dir
|
|
|
+ return os.path.join(path, dir, file)
|
|
|
+
|
|
|
+
|
|
|
+def filename_to_hex(filename):
|
|
|
+ """Takes an object filename and returns its corresponding hex sha."""
|
|
|
+ # grab the last (up to) two path components
|
|
|
+ names = filename.rsplit(os.path.sep, 2)[-2:]
|
|
|
+ errmsg = "Invalid object filename: %s" % filename
|
|
|
+ assert len(names) == 2, errmsg
|
|
|
+ base, rest = names
|
|
|
+ assert len(base) == 2 and len(rest) == 38, errmsg
|
|
|
+ hex = base + rest
|
|
|
+ hex_to_sha(hex)
|
|
|
+ return hex
|
|
|
+
|
|
|
+
|
|
|
def serializable_property(name, docstring=None):
|
|
|
def set(obj, value):
|
|
|
obj._ensure_parsed()
|
|
@@ -89,44 +117,100 @@ def serializable_property(name, docstring=None):
|
|
|
return property(get, set, doc=docstring)
|
|
|
|
|
|
|
|
|
+def object_class(type):
|
|
|
+ """Get the object class corresponding to the given type.
|
|
|
+
|
|
|
+ :param type: Either a type name string or a numeric type.
|
|
|
+ :return: The ShaFile subclass corresponding to the given type, or None if
|
|
|
+ type is not a valid type name/number.
|
|
|
+ """
|
|
|
+ return _TYPE_MAP.get(type, None)
|
|
|
+
|
|
|
+
|
|
|
+def check_hexsha(hex, error_msg):
|
|
|
+ try:
|
|
|
+ hex_to_sha(hex)
|
|
|
+ except (TypeError, AssertionError):
|
|
|
+ raise ObjectFormatException("%s %s" % (error_msg, hex))
|
|
|
+
|
|
|
+
|
|
|
+def check_identity(identity, error_msg):
|
|
|
+ email_start = identity.find("<")
|
|
|
+ email_end = identity.find(">")
|
|
|
+ if (email_start < 0 or email_end < 0 or email_end <= email_start
|
|
|
+ or identity.find("<", email_start + 1) >= 0
|
|
|
+ or identity.find(">", email_end + 1) >= 0
|
|
|
+ or not identity.endswith(">")):
|
|
|
+ raise ObjectFormatException(error_msg)
|
|
|
+
|
|
|
+
|
|
|
+class FixedSha(object):
|
|
|
+ """SHA object that behaves like hashlib's but is given a fixed value."""
|
|
|
+
|
|
|
+ def __init__(self, hexsha):
|
|
|
+ self._hexsha = hexsha
|
|
|
+ self._sha = hex_to_sha(hexsha)
|
|
|
+
|
|
|
+ def digest(self):
|
|
|
+ return self._sha
|
|
|
+
|
|
|
+ def hexdigest(self):
|
|
|
+ return self._hexsha
|
|
|
+
|
|
|
+
|
|
|
class ShaFile(object):
|
|
|
"""A git SHA file."""
|
|
|
|
|
|
- @classmethod
|
|
|
- def _parse_legacy_object(cls, map):
|
|
|
- """Parse a legacy object, creating it and setting object._text"""
|
|
|
- text = _decompress(map)
|
|
|
- object = None
|
|
|
- for posstype in type_map.keys():
|
|
|
- if text.startswith(posstype):
|
|
|
- object = type_map[posstype]()
|
|
|
- text = text[len(posstype):]
|
|
|
- break
|
|
|
- assert object is not None, "%s is not a known object type" % text[:9]
|
|
|
- assert text[0] == ' ', "%s is not a space" % text[0]
|
|
|
- text = text[1:]
|
|
|
- size = 0
|
|
|
- i = 0
|
|
|
- while text[0] >= '0' and text[0] <= '9':
|
|
|
- if i > 0 and size == 0:
|
|
|
- raise AssertionError("Size is not in canonical format")
|
|
|
- size = (size * 10) + int(text[0])
|
|
|
- text = text[1:]
|
|
|
- i += 1
|
|
|
- object._size = size
|
|
|
- assert text[0] == "\0", "Size not followed by null"
|
|
|
- text = text[1:]
|
|
|
- object.set_raw_string(text)
|
|
|
- return object
|
|
|
+ @staticmethod
|
|
|
+ def _parse_legacy_object_header(magic, f):
|
|
|
+ """Parse a legacy object, creating it but not reading the file."""
|
|
|
+ bufsize = 1024
|
|
|
+ decomp = zlib.decompressobj()
|
|
|
+ header = decomp.decompress(magic)
|
|
|
+ start = 0
|
|
|
+ end = -1
|
|
|
+ while end < 0:
|
|
|
+ header += decomp.decompress(f.read(bufsize))
|
|
|
+ end = header.find("\0", start)
|
|
|
+ start = len(header)
|
|
|
+ header = header[:end]
|
|
|
+ type_name, size = header.split(" ", 1)
|
|
|
+ size = int(size) # sanity check
|
|
|
+ obj_class = object_class(type_name)
|
|
|
+ if not obj_class:
|
|
|
+ raise ObjectFormatException("Not a known type: %s" % type_name)
|
|
|
+ obj = obj_class()
|
|
|
+ obj._filename = f.name
|
|
|
+ return obj
|
|
|
+
|
|
|
+ def _parse_legacy_object(self, f):
|
|
|
+ """Parse a legacy object, setting the raw string."""
|
|
|
+ size = os.path.getsize(f.name)
|
|
|
+ map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
|
|
|
+ try:
|
|
|
+ text = _decompress(map)
|
|
|
+ finally:
|
|
|
+ map.close()
|
|
|
+ header_end = text.find('\0')
|
|
|
+ if header_end < 0:
|
|
|
+ raise ObjectFormatException("Invalid object header")
|
|
|
+ self.set_raw_string(text[header_end+1:])
|
|
|
+
|
|
|
+ def as_legacy_object_chunks(self):
|
|
|
+ compobj = zlib.compressobj()
|
|
|
+ yield compobj.compress(self._header())
|
|
|
+ for chunk in self.as_raw_chunks():
|
|
|
+ yield compobj.compress(chunk)
|
|
|
+ yield compobj.flush()
|
|
|
|
|
|
def as_legacy_object(self):
|
|
|
- text = self.as_raw_string()
|
|
|
- return zlib.compress("%s %d\0%s" % (self._type, len(text), text))
|
|
|
+ return "".join(self.as_legacy_object_chunks())
|
|
|
|
|
|
def as_raw_chunks(self):
|
|
|
- if self._needs_serialization:
|
|
|
+ if self._needs_parsing:
|
|
|
+ self._ensure_parsed()
|
|
|
+ elif self._needs_serialization:
|
|
|
self._chunked_text = self._serialize()
|
|
|
- self._needs_serialization = False
|
|
|
return self._chunked_text
|
|
|
|
|
|
def as_raw_string(self):
|
|
@@ -143,6 +227,9 @@ class ShaFile(object):
|
|
|
|
|
|
def _ensure_parsed(self):
|
|
|
if self._needs_parsing:
|
|
|
+ if not self._chunked_text:
|
|
|
+ assert self._filename, "ShaFile needs either text or filename"
|
|
|
+ self._parse_file()
|
|
|
self._deserialize(self._chunked_text)
|
|
|
self._needs_parsing = False
|
|
|
|
|
@@ -153,39 +240,60 @@ class ShaFile(object):
|
|
|
|
|
|
def set_raw_chunks(self, chunks):
|
|
|
self._chunked_text = chunks
|
|
|
+ self._deserialize(chunks)
|
|
|
self._sha = None
|
|
|
- self._needs_parsing = True
|
|
|
+ self._needs_parsing = False
|
|
|
self._needs_serialization = False
|
|
|
|
|
|
- @classmethod
|
|
|
- def _parse_object(cls, map):
|
|
|
- """Parse a new style object , creating it and setting object._text"""
|
|
|
- used = 0
|
|
|
- byte = ord(map[used])
|
|
|
- used += 1
|
|
|
- num_type = (byte >> 4) & 7
|
|
|
+ @staticmethod
|
|
|
+ def _parse_object_header(magic, f):
|
|
|
+ """Parse a new style object, creating it but not reading the file."""
|
|
|
+ num_type = (ord(magic[0]) >> 4) & 7
|
|
|
+ obj_class = object_class(num_type)
|
|
|
+ if not obj_class:
|
|
|
+ raise ObjectFormatError("Not a known type: %d" % num_type)
|
|
|
+ obj = obj_class()
|
|
|
+ obj._filename = f.name
|
|
|
+ return obj
|
|
|
+
|
|
|
+ def _parse_object(self, f):
|
|
|
+ """Parse a new style object, setting self._text."""
|
|
|
+ size = os.path.getsize(f.name)
|
|
|
+ map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
|
|
|
try:
|
|
|
- object = num_type_map[num_type]()
|
|
|
- except KeyError:
|
|
|
- raise AssertionError("Not a known type: %d" % num_type)
|
|
|
- while (byte & 0x80) != 0:
|
|
|
- byte = ord(map[used])
|
|
|
- used += 1
|
|
|
- raw = map[used:]
|
|
|
- object.set_raw_string(_decompress(raw))
|
|
|
- return object
|
|
|
+ # skip type and size; type must have already been determined, and we
|
|
|
+ # trust zlib to fail if it's otherwise corrupted
|
|
|
+ byte = ord(map[0])
|
|
|
+ used = 1
|
|
|
+ while (byte & 0x80) != 0:
|
|
|
+ byte = ord(map[used])
|
|
|
+ used += 1
|
|
|
+ raw = map[used:]
|
|
|
+ self.set_raw_string(_decompress(raw))
|
|
|
+ finally:
|
|
|
+ map.close()
|
|
|
+
|
|
|
+ @classmethod
|
|
|
+ def _is_legacy_object(cls, magic):
|
|
|
+ b0, b1 = map(ord, magic)
|
|
|
+ word = (b0 << 8) + b1
|
|
|
+ return b0 == 0x78 and (word % 31) == 0
|
|
|
|
|
|
@classmethod
|
|
|
- def _parse_file(cls, map):
|
|
|
- word = (ord(map[0]) << 8) + ord(map[1])
|
|
|
- if ord(map[0]) == 0x78 and (word % 31) == 0:
|
|
|
- return cls._parse_legacy_object(map)
|
|
|
+ def _parse_file_header(cls, f):
|
|
|
+ magic = f.read(2)
|
|
|
+ if cls._is_legacy_object(magic):
|
|
|
+ return cls._parse_legacy_object_header(magic, f)
|
|
|
else:
|
|
|
- return cls._parse_object(map)
|
|
|
+ return cls._parse_object_header(magic, f)
|
|
|
|
|
|
def __init__(self):
|
|
|
"""Don't call this directly"""
|
|
|
self._sha = None
|
|
|
+ self._filename = None
|
|
|
+ self._chunked_text = []
|
|
|
+ self._needs_parsing = False
|
|
|
+ self._needs_serialization = True
|
|
|
|
|
|
def _deserialize(self, chunks):
|
|
|
raise NotImplementedError(self._deserialize)
|
|
@@ -193,53 +301,96 @@ class ShaFile(object):
|
|
|
def _serialize(self):
|
|
|
raise NotImplementedError(self._serialize)
|
|
|
|
|
|
+ def _parse_file(self):
|
|
|
+ f = GitFile(self._filename, 'rb')
|
|
|
+ try:
|
|
|
+ magic = f.read(2)
|
|
|
+ if self._is_legacy_object(magic):
|
|
|
+ self._parse_legacy_object(f)
|
|
|
+ else:
|
|
|
+ self._parse_object(f)
|
|
|
+ finally:
|
|
|
+ f.close()
|
|
|
+
|
|
|
@classmethod
|
|
|
def from_file(cls, filename):
|
|
|
- """Get the contents of a SHA file on disk"""
|
|
|
- size = os.path.getsize(filename)
|
|
|
+ """Get the contents of a SHA file on disk."""
|
|
|
f = GitFile(filename, 'rb')
|
|
|
try:
|
|
|
- map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
|
|
|
- shafile = cls._parse_file(map)
|
|
|
- return shafile
|
|
|
+ try:
|
|
|
+ obj = cls._parse_file_header(f)
|
|
|
+ obj._sha = FixedSha(filename_to_hex(filename))
|
|
|
+ obj._needs_parsing = True
|
|
|
+ obj._needs_serialization = True
|
|
|
+ return obj
|
|
|
+ except (IndexError, ValueError), e:
|
|
|
+ raise ObjectFormatException("invalid object header")
|
|
|
finally:
|
|
|
f.close()
|
|
|
|
|
|
- @classmethod
|
|
|
- def from_raw_string(cls, type, string):
|
|
|
+ @staticmethod
|
|
|
+ def from_raw_string(type_num, string):
|
|
|
"""Creates an object of the indicated type from the raw string given.
|
|
|
|
|
|
- Type is the numeric type of an object. String is the raw uncompressed
|
|
|
- contents.
|
|
|
+ :param type_num: The numeric type of the object.
|
|
|
+ :param string: The raw uncompressed contents.
|
|
|
"""
|
|
|
- real_class = num_type_map[type]
|
|
|
- obj = real_class()
|
|
|
- obj.type = type
|
|
|
+ obj = object_class(type_num)()
|
|
|
obj.set_raw_string(string)
|
|
|
return obj
|
|
|
|
|
|
- @classmethod
|
|
|
- def from_raw_chunks(cls, type, chunks):
|
|
|
+ @staticmethod
|
|
|
+ def from_raw_chunks(type_num, chunks):
|
|
|
"""Creates an object of the indicated type from the raw chunks given.
|
|
|
|
|
|
- Type is the numeric type of an object. Chunks is a sequence of the raw
|
|
|
- uncompressed contents.
|
|
|
+ :param type_num: The numeric type of the object.
|
|
|
+ :param chunks: An iterable of the raw uncompressed contents.
|
|
|
"""
|
|
|
- real_class = num_type_map[type]
|
|
|
- obj = real_class()
|
|
|
- obj.type = type
|
|
|
+ obj = object_class(type_num)()
|
|
|
obj.set_raw_chunks(chunks)
|
|
|
return obj
|
|
|
|
|
|
@classmethod
|
|
|
def from_string(cls, string):
|
|
|
- """Create a blob from a string."""
|
|
|
- shafile = cls()
|
|
|
- shafile.set_raw_string(string)
|
|
|
- return shafile
|
|
|
+ """Create a ShaFile from a string."""
|
|
|
+ obj = cls()
|
|
|
+ obj.set_raw_string(string)
|
|
|
+ return obj
|
|
|
+
|
|
|
+ def _check_has_member(self, member, error_msg):
|
|
|
+ """Check that the object has a given member variable.
|
|
|
+
|
|
|
+ :param member: the member variable to check for
|
|
|
+ :param error_msg: the message for an error if the member is missing
|
|
|
+ :raise ObjectFormatException: with the given error_msg if member is
|
|
|
+ missing or is None
|
|
|
+ """
|
|
|
+ if getattr(self, member, None) is None:
|
|
|
+ raise ObjectFormatException(error_msg)
|
|
|
+
|
|
|
+ def check(self):
|
|
|
+ """Check this object for internal consistency.
|
|
|
+
|
|
|
+ :raise ObjectFormatException: if the object is malformed in some way
|
|
|
+ :raise ChecksumMismatch: if the object was created with a SHA that does
|
|
|
+ not match its contents
|
|
|
+ """
|
|
|
+ # TODO: if we find that error-checking during object parsing is a
|
|
|
+ # performance bottleneck, those checks should be moved to the class's
|
|
|
+ # check() method during optimization so we can still check the object
|
|
|
+ # when necessary.
|
|
|
+ old_sha = self.id
|
|
|
+ try:
|
|
|
+ self._deserialize(self.as_raw_chunks())
|
|
|
+ self._sha = None
|
|
|
+ new_sha = self.id
|
|
|
+ except Exception, e:
|
|
|
+ raise ObjectFormatException(e)
|
|
|
+ if old_sha != new_sha:
|
|
|
+ raise ChecksumMismatch(new_sha, old_sha)
|
|
|
|
|
|
def _header(self):
|
|
|
- return "%s %lu\0" % (self._type, self.raw_length())
|
|
|
+ return "%s %lu\0" % (self.type_name, self.raw_length())
|
|
|
|
|
|
def raw_length(self):
|
|
|
"""Returns the length of the raw string of this object."""
|
|
@@ -257,8 +408,13 @@ class ShaFile(object):
|
|
|
|
|
|
def sha(self):
|
|
|
"""The SHA1 object that is the name of this object."""
|
|
|
- if self._needs_serialization or self._sha is None:
|
|
|
- self._sha = self._make_sha()
|
|
|
+ if self._sha is None:
|
|
|
+ # this is a local because as_raw_chunks() overwrites self._sha
|
|
|
+ new_sha = make_sha()
|
|
|
+ new_sha.update(self._header())
|
|
|
+ for chunk in self.as_raw_chunks():
|
|
|
+ new_sha.update(chunk)
|
|
|
+ self._sha = new_sha
|
|
|
return self._sha
|
|
|
|
|
|
@property
|
|
@@ -266,11 +422,12 @@ class ShaFile(object):
|
|
|
return self.sha().hexdigest()
|
|
|
|
|
|
def get_type(self):
|
|
|
- return self._num_type
|
|
|
+ return self.type_num
|
|
|
|
|
|
def set_type(self, type):
|
|
|
- self._num_type = type
|
|
|
+ self.type_num = type
|
|
|
|
|
|
+ # DEPRECATED: use type_num or type_name as needed.
|
|
|
type = property(get_type, set_type)
|
|
|
|
|
|
def __repr__(self):
|
|
@@ -291,8 +448,8 @@ class ShaFile(object):
|
|
|
class Blob(ShaFile):
|
|
|
"""A Git Blob object."""
|
|
|
|
|
|
- _type = BLOB_ID
|
|
|
- _num_type = 3
|
|
|
+ type_name = 'blob'
|
|
|
+ type_num = 3
|
|
|
|
|
|
def __init__(self):
|
|
|
super(Blob, self).__init__()
|
|
@@ -307,60 +464,125 @@ class Blob(ShaFile):
|
|
|
self.set_raw_string(data)
|
|
|
|
|
|
data = property(_get_data, _set_data,
|
|
|
- "The text contained within the blob object.")
|
|
|
+ "The text contained within the blob object.")
|
|
|
|
|
|
def _get_chunked(self):
|
|
|
+ self._ensure_parsed()
|
|
|
return self._chunked_text
|
|
|
|
|
|
def _set_chunked(self, chunks):
|
|
|
self._chunked_text = chunks
|
|
|
|
|
|
+ def _serialize(self):
|
|
|
+ if not self._chunked_text:
|
|
|
+ self._ensure_parsed()
|
|
|
+ self._needs_serialization = False
|
|
|
+ return self._chunked_text
|
|
|
+
|
|
|
+ def _deserialize(self, chunks):
|
|
|
+ self._chunked_text = chunks
|
|
|
+
|
|
|
chunked = property(_get_chunked, _set_chunked,
|
|
|
"The text within the blob object, as chunks (not necessarily lines).")
|
|
|
|
|
|
@classmethod
|
|
|
def from_file(cls, filename):
|
|
|
blob = ShaFile.from_file(filename)
|
|
|
- if blob._type != cls._type:
|
|
|
+ if not isinstance(blob, cls):
|
|
|
raise NotBlobError(filename)
|
|
|
return blob
|
|
|
|
|
|
+ def check(self):
|
|
|
+ """Check this object for internal consistency.
|
|
|
+
|
|
|
+ :raise ObjectFormatException: if the object is malformed in some way
|
|
|
+ """
|
|
|
+ super(Blob, self).check()
|
|
|
+
|
|
|
+
|
|
|
+def _parse_tag_or_commit(text):
|
|
|
+ """Parse tag or commit text.
|
|
|
+
|
|
|
+ :param text: the raw text of the tag or commit object.
|
|
|
+ :yield: tuples of (field, value), one per header line, in the order read
|
|
|
+ from the text, possibly including duplicates. Includes a field named
|
|
|
+ None for the freeform tag/commit text.
|
|
|
+ """
|
|
|
+ f = StringIO(text)
|
|
|
+ for l in f:
|
|
|
+ l = l.rstrip("\n")
|
|
|
+ if l == "":
|
|
|
+ # Empty line indicates end of headers
|
|
|
+ break
|
|
|
+ yield l.split(" ", 1)
|
|
|
+ yield (None, f.read())
|
|
|
+ f.close()
|
|
|
+
|
|
|
+
|
|
|
+def parse_tag(text):
|
|
|
+ return _parse_tag_or_commit(text)
|
|
|
+
|
|
|
|
|
|
class Tag(ShaFile):
|
|
|
"""A Git Tag object."""
|
|
|
|
|
|
- _type = TAG_ID
|
|
|
- _num_type = 4
|
|
|
+ type_name = 'tag'
|
|
|
+ type_num = 4
|
|
|
|
|
|
def __init__(self):
|
|
|
super(Tag, self).__init__()
|
|
|
- self._needs_parsing = False
|
|
|
- self._needs_serialization = True
|
|
|
+ self._tag_timezone_neg_utc = False
|
|
|
|
|
|
@classmethod
|
|
|
def from_file(cls, filename):
|
|
|
- blob = ShaFile.from_file(filename)
|
|
|
- if blob._type != cls._type:
|
|
|
- raise NotBlobError(filename)
|
|
|
- return blob
|
|
|
+ tag = ShaFile.from_file(filename)
|
|
|
+ if not isinstance(tag, cls):
|
|
|
+ raise NotTagError(filename)
|
|
|
+ return tag
|
|
|
|
|
|
- @classmethod
|
|
|
- def from_string(cls, string):
|
|
|
- """Create a blob from a string."""
|
|
|
- shafile = cls()
|
|
|
- shafile.set_raw_string(string)
|
|
|
- return shafile
|
|
|
+ def check(self):
|
|
|
+ """Check this object for internal consistency.
|
|
|
+
|
|
|
+ :raise ObjectFormatException: if the object is malformed in some way
|
|
|
+ """
|
|
|
+ super(Tag, self).check()
|
|
|
+ self._check_has_member("_object_sha", "missing object sha")
|
|
|
+ self._check_has_member("_object_class", "missing object type")
|
|
|
+ self._check_has_member("_name", "missing tag name")
|
|
|
+
|
|
|
+ if not self._name:
|
|
|
+ raise ObjectFormatException("empty tag name")
|
|
|
+
|
|
|
+ check_hexsha(self._object_sha, "invalid object sha")
|
|
|
+
|
|
|
+ if getattr(self, "_tagger", None):
|
|
|
+ check_identity(self._tagger, "invalid tagger")
|
|
|
+
|
|
|
+ last = None
|
|
|
+ for field, _ in parse_tag("".join(self._chunked_text)):
|
|
|
+ if field == _OBJECT_HEADER and last is not None:
|
|
|
+ raise ObjectFormatException("unexpected object")
|
|
|
+ elif field == _TYPE_HEADER and last != _OBJECT_HEADER:
|
|
|
+ raise ObjectFormatException("unexpected type")
|
|
|
+ elif field == _TAG_HEADER and last != _TYPE_HEADER:
|
|
|
+ raise ObjectFormatException("unexpected tag name")
|
|
|
+ elif field == _TAGGER_HEADER and last != _TAG_HEADER:
|
|
|
+ raise ObjectFormatException("unexpected tagger")
|
|
|
+ last = field
|
|
|
|
|
|
def _serialize(self):
|
|
|
chunks = []
|
|
|
- chunks.append("%s %s\n" % (OBJECT_ID, self._object_sha))
|
|
|
- chunks.append("%s %s\n" % (TYPE_ID, num_type_map[self._object_type]._type))
|
|
|
- chunks.append("%s %s\n" % (TAG_ID, self._name))
|
|
|
+ chunks.append("%s %s\n" % (_OBJECT_HEADER, self._object_sha))
|
|
|
+ chunks.append("%s %s\n" % (_TYPE_HEADER, self._object_class.type_name))
|
|
|
+ chunks.append("%s %s\n" % (_TAG_HEADER, self._name))
|
|
|
if self._tagger:
|
|
|
if self._tag_time is None:
|
|
|
- chunks.append("%s %s\n" % (TAGGER_ID, self._tagger))
|
|
|
+ chunks.append("%s %s\n" % (_TAGGER_HEADER, self._tagger))
|
|
|
else:
|
|
|
- chunks.append("%s %s %d %s\n" % (TAGGER_ID, self._tagger, self._tag_time, format_timezone(self._tag_timezone)))
|
|
|
+ chunks.append("%s %s %d %s\n" % (
|
|
|
+ _TAGGER_HEADER, self._tagger, self._tag_time,
|
|
|
+ format_timezone(self._tag_timezone,
|
|
|
+ self._tag_timezone_neg_utc)))
|
|
|
chunks.append("\n") # To close headers
|
|
|
chunks.append(self._message)
|
|
|
return chunks
|
|
@@ -368,45 +590,49 @@ class Tag(ShaFile):
|
|
|
def _deserialize(self, chunks):
|
|
|
"""Grab the metadata attached to the tag"""
|
|
|
self._tagger = None
|
|
|
- f = StringIO("".join(chunks))
|
|
|
- for l in f:
|
|
|
- l = l.rstrip("\n")
|
|
|
- if l == "":
|
|
|
- break # empty line indicates end of headers
|
|
|
- (field, value) = l.split(" ", 1)
|
|
|
- if field == OBJECT_ID:
|
|
|
+ for field, value in parse_tag("".join(chunks)):
|
|
|
+ if field == _OBJECT_HEADER:
|
|
|
self._object_sha = value
|
|
|
- elif field == TYPE_ID:
|
|
|
- self._object_type = type_map[value]
|
|
|
- elif field == TAG_ID:
|
|
|
+ elif field == _TYPE_HEADER:
|
|
|
+ obj_class = object_class(value)
|
|
|
+ if not obj_class:
|
|
|
+ raise ObjectFormatException("Not a known type: %s" % value)
|
|
|
+ self._object_class = obj_class
|
|
|
+ elif field == _TAG_HEADER:
|
|
|
self._name = value
|
|
|
- elif field == TAGGER_ID:
|
|
|
+ elif field == _TAGGER_HEADER:
|
|
|
try:
|
|
|
sep = value.index("> ")
|
|
|
except ValueError:
|
|
|
self._tagger = value
|
|
|
self._tag_time = None
|
|
|
self._tag_timezone = None
|
|
|
+ self._tag_timezone_neg_utc = False
|
|
|
else:
|
|
|
self._tagger = value[0:sep+1]
|
|
|
- (timetext, timezonetext) = value[sep+2:].rsplit(" ", 1)
|
|
|
try:
|
|
|
+ (timetext, timezonetext) = value[sep+2:].rsplit(" ", 1)
|
|
|
self._tag_time = int(timetext)
|
|
|
- except ValueError: #Not a unix timestamp
|
|
|
- self._tag_time = time.strptime(timetext)
|
|
|
- self._tag_timezone = parse_timezone(timezonetext)
|
|
|
+ self._tag_timezone, self._tag_timezone_neg_utc = \
|
|
|
+ parse_timezone(timezonetext)
|
|
|
+ except ValueError, e:
|
|
|
+ raise ObjectFormatException(e)
|
|
|
+ elif field is None:
|
|
|
+ self._message = value
|
|
|
else:
|
|
|
- raise AssertionError("Unknown field %s" % field)
|
|
|
- self._message = f.read()
|
|
|
+ raise ObjectFormatError("Unknown field %s" % field)
|
|
|
|
|
|
def _get_object(self):
|
|
|
- """Returns the object pointed by this tag, represented as a tuple(type, sha)"""
|
|
|
+ """Get the object pointed to by this tag.
|
|
|
+
|
|
|
+ :return: tuple of (object class, sha).
|
|
|
+ """
|
|
|
self._ensure_parsed()
|
|
|
- return (self._object_type, self._object_sha)
|
|
|
+ return (self._object_class, self._object_sha)
|
|
|
|
|
|
def _set_object(self, value):
|
|
|
self._ensure_parsed()
|
|
|
- (self._object_type, self._object_sha) = value
|
|
|
+ (self._object_class, self._object_sha) = value
|
|
|
self._needs_serialization = True
|
|
|
|
|
|
object = property(_get_object, _set_object)
|
|
@@ -425,9 +651,8 @@ def parse_tree(text):
|
|
|
"""Parse a tree text.
|
|
|
|
|
|
:param text: Serialized text to parse
|
|
|
- :return: Dictionary with names as keys, (mode, sha) tuples as values
|
|
|
+ :yields: tuples of (name, mode, sha)
|
|
|
"""
|
|
|
- ret = {}
|
|
|
count = 0
|
|
|
l = len(text)
|
|
|
while count < l:
|
|
@@ -437,8 +662,7 @@ def parse_tree(text):
|
|
|
name = text[mode_end+1:name_end]
|
|
|
count = name_end+21
|
|
|
sha = text[name_end+1:count]
|
|
|
- ret[name] = (mode, sha_to_hex(sha))
|
|
|
- return ret
|
|
|
+ yield (name, mode, sha_to_hex(sha))
|
|
|
|
|
|
|
|
|
def serialize_tree(items):
|
|
@@ -458,32 +682,33 @@ def sorted_tree_items(entries):
|
|
|
:param entries: Dictionary mapping names to (mode, sha) tuples
|
|
|
:return: Iterator over (name, mode, sha)
|
|
|
"""
|
|
|
- def cmp_entry((name1, value1), (name2, value2)):
|
|
|
- if stat.S_ISDIR(value1[0]):
|
|
|
- name1 += "/"
|
|
|
- if stat.S_ISDIR(value2[0]):
|
|
|
- name2 += "/"
|
|
|
- return cmp(name1, name2)
|
|
|
for name, entry in sorted(entries.iteritems(), cmp=cmp_entry):
|
|
|
yield name, entry[0], entry[1]
|
|
|
|
|
|
|
|
|
+def cmp_entry((name1, value1), (name2, value2)):
|
|
|
+ """Compare two tree entries."""
|
|
|
+ if stat.S_ISDIR(value1[0]):
|
|
|
+ name1 += "/"
|
|
|
+ if stat.S_ISDIR(value2[0]):
|
|
|
+ name2 += "/"
|
|
|
+ return cmp(name1, name2)
|
|
|
+
|
|
|
+
|
|
|
class Tree(ShaFile):
|
|
|
"""A Git tree object"""
|
|
|
|
|
|
- _type = TREE_ID
|
|
|
- _num_type = 2
|
|
|
+ type_name = 'tree'
|
|
|
+ type_num = 2
|
|
|
|
|
|
def __init__(self):
|
|
|
super(Tree, self).__init__()
|
|
|
self._entries = {}
|
|
|
- self._needs_parsing = False
|
|
|
- self._needs_serialization = True
|
|
|
|
|
|
@classmethod
|
|
|
def from_file(cls, filename):
|
|
|
tree = ShaFile.from_file(filename)
|
|
|
- if tree._type != cls._type:
|
|
|
+ if not isinstance(tree, cls):
|
|
|
raise NotTreeError(filename)
|
|
|
return tree
|
|
|
|
|
@@ -511,6 +736,10 @@ class Tree(ShaFile):
|
|
|
self._ensure_parsed()
|
|
|
return len(self._entries)
|
|
|
|
|
|
+ def __iter__(self):
|
|
|
+ self._ensure_parsed()
|
|
|
+ return iter(self._entries)
|
|
|
+
|
|
|
def add(self, mode, name, hexsha):
|
|
|
assert type(mode) == int
|
|
|
assert type(name) == str
|
|
@@ -528,8 +757,7 @@ class Tree(ShaFile):
|
|
|
(mode, name, hexsha) for (name, mode, hexsha) in self.iteritems()]
|
|
|
|
|
|
def iteritems(self):
|
|
|
- """Iterate over all entries in the order in which they would be
|
|
|
- serialized.
|
|
|
+ """Iterate over entries in the order in which they would be serialized.
|
|
|
|
|
|
:return: Iterator over (name, mode, sha) tuples
|
|
|
"""
|
|
@@ -538,7 +766,40 @@ class Tree(ShaFile):
|
|
|
|
|
|
def _deserialize(self, chunks):
|
|
|
"""Grab the entries in the tree"""
|
|
|
- self._entries = parse_tree("".join(chunks))
|
|
|
+ try:
|
|
|
+ parsed_entries = parse_tree("".join(chunks))
|
|
|
+ except ValueError, e:
|
|
|
+ raise ObjectFormatException(e)
|
|
|
+ # TODO: list comprehension is for efficiency in the common (small) case;
|
|
|
+ # if memory efficiency in the large case is a concern, use a genexp.
|
|
|
+ self._entries = dict([(n, (m, s)) for n, m, s in parsed_entries])
|
|
|
+
|
|
|
+ def check(self):
|
|
|
+ """Check this object for internal consistency.
|
|
|
+
|
|
|
+ :raise ObjectFormatException: if the object is malformed in some way
|
|
|
+ """
|
|
|
+ super(Tree, self).check()
|
|
|
+ last = None
|
|
|
+ allowed_modes = (stat.S_IFREG | 0755, stat.S_IFREG | 0644,
|
|
|
+ stat.S_IFLNK, stat.S_IFDIR, S_IFGITLINK,
|
|
|
+ # TODO: optionally exclude as in git fsck --strict
|
|
|
+ stat.S_IFREG | 0664)
|
|
|
+ for name, mode, sha in parse_tree("".join(self._chunked_text)):
|
|
|
+ check_hexsha(sha, 'invalid sha %s' % sha)
|
|
|
+ if '/' in name or name in ('', '.', '..'):
|
|
|
+ raise ObjectFormatException('invalid name %s' % name)
|
|
|
+
|
|
|
+ if mode not in allowed_modes:
|
|
|
+ raise ObjectFormatException('invalid mode %06o' % mode)
|
|
|
+
|
|
|
+ entry = (name, (mode, sha))
|
|
|
+ if last:
|
|
|
+ if cmp_entry(last, entry) > 0:
|
|
|
+ raise ObjectFormatException('entries not sorted')
|
|
|
+ if name == last[0]:
|
|
|
+ raise ObjectFormatException('duplicate entry %s' % name)
|
|
|
+ last = entry
|
|
|
|
|
|
def _serialize(self):
|
|
|
return list(serialize_tree(self.iteritems()))
|
|
@@ -556,39 +817,47 @@ class Tree(ShaFile):
|
|
|
|
|
|
def parse_timezone(text):
|
|
|
offset = int(text)
|
|
|
+ negative_utc = (offset == 0 and text[0] == '-')
|
|
|
signum = (offset < 0) and -1 or 1
|
|
|
offset = abs(offset)
|
|
|
hours = int(offset / 100)
|
|
|
minutes = (offset % 100)
|
|
|
- return signum * (hours * 3600 + minutes * 60)
|
|
|
+ return signum * (hours * 3600 + minutes * 60), negative_utc
|
|
|
|
|
|
|
|
|
-def format_timezone(offset):
|
|
|
+def format_timezone(offset, negative_utc=False):
|
|
|
if offset % 60 != 0:
|
|
|
raise ValueError("Unable to handle non-minute offset.")
|
|
|
- sign = (offset < 0) and '-' or '+'
|
|
|
+ if offset < 0 or (offset == 0 and negative_utc):
|
|
|
+ sign = '-'
|
|
|
+ else:
|
|
|
+ sign = '+'
|
|
|
offset = abs(offset)
|
|
|
return '%c%02d%02d' % (sign, offset / 3600, (offset / 60) % 60)
|
|
|
|
|
|
|
|
|
+def parse_commit(text):
|
|
|
+ return _parse_tag_or_commit(text)
|
|
|
+
|
|
|
+
|
|
|
class Commit(ShaFile):
|
|
|
"""A git commit object"""
|
|
|
|
|
|
- _type = COMMIT_ID
|
|
|
- _num_type = 1
|
|
|
+ type_name = 'commit'
|
|
|
+ type_num = 1
|
|
|
|
|
|
def __init__(self):
|
|
|
super(Commit, self).__init__()
|
|
|
self._parents = []
|
|
|
self._encoding = None
|
|
|
- self._needs_parsing = False
|
|
|
- self._needs_serialization = True
|
|
|
self._extra = {}
|
|
|
+ self._author_timezone_neg_utc = False
|
|
|
+ self._commit_timezone_neg_utc = False
|
|
|
|
|
|
@classmethod
|
|
|
def from_file(cls, filename):
|
|
|
commit = ShaFile.from_file(filename)
|
|
|
- if commit._type != cls._type:
|
|
|
+ if not isinstance(commit, cls):
|
|
|
raise NotCommitError(filename)
|
|
|
return commit
|
|
|
|
|
@@ -596,40 +865,79 @@ class Commit(ShaFile):
|
|
|
self._parents = []
|
|
|
self._extra = []
|
|
|
self._author = None
|
|
|
- f = StringIO("".join(chunks))
|
|
|
- for l in f:
|
|
|
- l = l.rstrip("\n")
|
|
|
- if l == "":
|
|
|
- # Empty line indicates end of headers
|
|
|
- break
|
|
|
- (field, value) = l.split(" ", 1)
|
|
|
- if field == TREE_ID:
|
|
|
+ for field, value in parse_commit("".join(self._chunked_text)):
|
|
|
+ if field == _TREE_HEADER:
|
|
|
self._tree = value
|
|
|
- elif field == PARENT_ID:
|
|
|
+ elif field == _PARENT_HEADER:
|
|
|
self._parents.append(value)
|
|
|
- elif field == AUTHOR_ID:
|
|
|
+ elif field == _AUTHOR_HEADER:
|
|
|
self._author, timetext, timezonetext = value.rsplit(" ", 2)
|
|
|
self._author_time = int(timetext)
|
|
|
- self._author_timezone = parse_timezone(timezonetext)
|
|
|
- elif field == COMMITTER_ID:
|
|
|
+ self._author_timezone, self._author_timezone_neg_utc =\
|
|
|
+ parse_timezone(timezonetext)
|
|
|
+ elif field == _COMMITTER_HEADER:
|
|
|
self._committer, timetext, timezonetext = value.rsplit(" ", 2)
|
|
|
self._commit_time = int(timetext)
|
|
|
- self._commit_timezone = parse_timezone(timezonetext)
|
|
|
- elif field == ENCODING_ID:
|
|
|
+ self._commit_timezone, self._commit_timezone_neg_utc =\
|
|
|
+ parse_timezone(timezonetext)
|
|
|
+ elif field == _ENCODING_HEADER:
|
|
|
self._encoding = value
|
|
|
+ elif field is None:
|
|
|
+ self._message = value
|
|
|
else:
|
|
|
self._extra.append((field, value))
|
|
|
- self._message = f.read()
|
|
|
+
|
|
|
+ def check(self):
|
|
|
+ """Check this object for internal consistency.
|
|
|
+
|
|
|
+ :raise ObjectFormatException: if the object is malformed in some way
|
|
|
+ """
|
|
|
+ super(Commit, self).check()
|
|
|
+ self._check_has_member("_tree", "missing tree")
|
|
|
+ self._check_has_member("_author", "missing author")
|
|
|
+ self._check_has_member("_committer", "missing committer")
|
|
|
+ # times are currently checked when set
|
|
|
+
|
|
|
+ for parent in self._parents:
|
|
|
+ check_hexsha(parent, "invalid parent sha")
|
|
|
+ check_hexsha(self._tree, "invalid tree sha")
|
|
|
+
|
|
|
+ check_identity(self._author, "invalid author")
|
|
|
+ check_identity(self._committer, "invalid committer")
|
|
|
+
|
|
|
+ last = None
|
|
|
+ for field, _ in parse_commit("".join(self._chunked_text)):
|
|
|
+ if field == _TREE_HEADER and last is not None:
|
|
|
+ raise ObjectFormatException("unexpected tree")
|
|
|
+ elif field == _PARENT_HEADER and last not in (_PARENT_HEADER,
|
|
|
+ _TREE_HEADER):
|
|
|
+ raise ObjectFormatException("unexpected parent")
|
|
|
+ elif field == _AUTHOR_HEADER and last not in (_TREE_HEADER,
|
|
|
+ _PARENT_HEADER):
|
|
|
+ raise ObjectFormatException("unexpected author")
|
|
|
+ elif field == _COMMITTER_HEADER and last != _AUTHOR_HEADER:
|
|
|
+ raise ObjectFormatException("unexpected committer")
|
|
|
+ elif field == _ENCODING_HEADER and last != _COMMITTER_HEADER:
|
|
|
+ raise ObjectFormatException("unexpected encoding")
|
|
|
+ last = field
|
|
|
+
|
|
|
+ # TODO: optionally check for duplicate parents
|
|
|
|
|
|
def _serialize(self):
|
|
|
chunks = []
|
|
|
- chunks.append("%s %s\n" % (TREE_ID, self._tree))
|
|
|
+ chunks.append("%s %s\n" % (_TREE_HEADER, self._tree))
|
|
|
for p in self._parents:
|
|
|
- chunks.append("%s %s\n" % (PARENT_ID, p))
|
|
|
- chunks.append("%s %s %s %s\n" % (AUTHOR_ID, self._author, str(self._author_time), format_timezone(self._author_timezone)))
|
|
|
- chunks.append("%s %s %s %s\n" % (COMMITTER_ID, self._committer, str(self._commit_time), format_timezone(self._commit_timezone)))
|
|
|
+ chunks.append("%s %s\n" % (_PARENT_HEADER, p))
|
|
|
+ chunks.append("%s %s %s %s\n" % (
|
|
|
+ _AUTHOR_HEADER, self._author, str(self._author_time),
|
|
|
+ format_timezone(self._author_timezone,
|
|
|
+ self._author_timezone_neg_utc)))
|
|
|
+ chunks.append("%s %s %s %s\n" % (
|
|
|
+ _COMMITTER_HEADER, self._committer, str(self._commit_time),
|
|
|
+ format_timezone(self._commit_timezone,
|
|
|
+ self._commit_timezone_neg_utc)))
|
|
|
if self.encoding:
|
|
|
- chunks.append("%s %s\n" % (ENCODING_ID, self.encoding))
|
|
|
+ chunks.append("%s %s\n" % (_ENCODING_HEADER, self.encoding))
|
|
|
for k, v in self.extra:
|
|
|
if "\n" in k or "\n" in v:
|
|
|
raise AssertionError("newline in extra data: %r -> %r" % (k, v))
|
|
@@ -685,22 +993,24 @@ class Commit(ShaFile):
|
|
|
"Encoding of the commit message.")
|
|
|
|
|
|
|
|
|
-type_map = {
|
|
|
- BLOB_ID : Blob,
|
|
|
- TREE_ID : Tree,
|
|
|
- COMMIT_ID : Commit,
|
|
|
- TAG_ID: Tag,
|
|
|
-}
|
|
|
+OBJECT_CLASSES = (
|
|
|
+ Commit,
|
|
|
+ Tree,
|
|
|
+ Blob,
|
|
|
+ Tag,
|
|
|
+ )
|
|
|
+
|
|
|
+_TYPE_MAP = {}
|
|
|
+
|
|
|
+for cls in OBJECT_CLASSES:
|
|
|
+ _TYPE_MAP[cls.type_name] = cls
|
|
|
+ _TYPE_MAP[cls.type_num] = cls
|
|
|
+
|
|
|
|
|
|
-num_type_map = {
|
|
|
- 0: None,
|
|
|
- 1: Commit,
|
|
|
- 2: Tree,
|
|
|
- 3: Blob,
|
|
|
- 4: Tag,
|
|
|
- # 5 Is reserved for further expansion
|
|
|
-}
|
|
|
|
|
|
+# Hold on to the pure-python implementations for testing
|
|
|
+_parse_tree_py = parse_tree
|
|
|
+_sorted_tree_items_py = sorted_tree_items
|
|
|
try:
|
|
|
# Try to import C versions
|
|
|
from dulwich._objects import parse_tree, sorted_tree_items
|