# objects.py -- Access to base git objects
# Copyright (C) 2007 James Westby <jw+debian@jameswestby.net>
-# Copyright (C) 2008 Jelmer Vernooij <jelmer@samba.org>
-#
+# Copyright (C) 2008-2009 Jelmer Vernooij <jelmer@samba.org>
+#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; version 2
# of the License or (at your option) a later version of the License.
-#
+#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
-#
+#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
"""Access to base git objects."""
+import binascii
+from cStringIO import (
+ StringIO,
+ )
import mmap
import os
-import sha
+import stat
import zlib
from dulwich.errors import (
+ ChecksumMismatch,
NotBlobError,
NotCommitError,
+ NotTagError,
NotTreeError,
+ ObjectFormatException,
+ )
+from dulwich.file import GitFile
+from dulwich.misc import (
+ make_sha,
)
-BLOB_ID = "blob"
-TAG_ID = "tag"
-TREE_ID = "tree"
-COMMIT_ID = "commit"
-PARENT_ID = "parent"
-AUTHOR_ID = "author"
-COMMITTER_ID = "committer"
-OBJECT_ID = "object"
-TYPE_ID = "type"
-TAGGER_ID = "tagger"
+
+# Header fields for commits
+_TREE_HEADER = "tree"
+_PARENT_HEADER = "parent"
+_AUTHOR_HEADER = "author"
+_COMMITTER_HEADER = "committer"
+_ENCODING_HEADER = "encoding"
+
+
+# Header fields for objects
+_OBJECT_HEADER = "object"
+_TYPE_HEADER = "type"
+_TAG_HEADER = "tag"
+_TAGGER_HEADER = "tagger"
+
+
+S_IFGITLINK = 0160000
+
+def S_ISGITLINK(m):
+ return (stat.S_IFMT(m) == S_IFGITLINK)
+
def _decompress(string):
dcomp = zlib.decompressobj()
def sha_to_hex(sha):
"""Takes a string and returns the hex of the sha within"""
- hexsha = "".join(["%02x" % ord(c) for c in sha])
+ hexsha = binascii.hexlify(sha)
assert len(hexsha) == 40, "Incorrect length of sha1 string: %d" % hexsha
return hexsha
def hex_to_sha(hex):
"""Takes a hex sha and returns a binary sha"""
assert len(hex) == 40, "Incorrent length of hexsha: %s" % hex
- return ''.join([chr(int(hex[i:i+2], 16)) for i in xrange(0, len(hex), 2)])
+ return binascii.unhexlify(hex)
+
+
+def hex_to_filename(path, hex):
+ """Takes a hex sha and returns its filename relative to the given path."""
+ dir = hex[:2]
+ file = hex[2:]
+ # Check from object dir
+ return os.path.join(path, dir, file)
+
+
+def filename_to_hex(filename):
+ """Takes an object filename and returns its corresponding hex sha."""
+ # grab the last (up to) two path components
+ names = filename.rsplit(os.path.sep, 2)[-2:]
+ errmsg = "Invalid object filename: %s" % filename
+ assert len(names) == 2, errmsg
+ base, rest = names
+ assert len(base) == 2 and len(rest) == 38, errmsg
+ hex = base + rest
+ hex_to_sha(hex)
+ return hex
+
+
+def object_header(num_type, length):
+ """Return an object header for the given numeric type and text length."""
+ return "%s %d\0" % (object_class(num_type).type_name, length)
+
+
+def serializable_property(name, docstring=None):
+ def set(obj, value):
+ obj._ensure_parsed()
+ setattr(obj, "_"+name, value)
+ obj._needs_serialization = True
+ def get(obj):
+ obj._ensure_parsed()
+ return getattr(obj, "_"+name)
+ return property(get, set, doc=docstring)
+
+
+def object_class(type):
+ """Get the object class corresponding to the given type.
+
+ :param type: Either a type name string or a numeric type.
+ :return: The ShaFile subclass corresponding to the given type, or None if
+ type is not a valid type name/number.
+ """
+ return _TYPE_MAP.get(type, None)
+
+
+def check_hexsha(hex, error_msg):
+ try:
+ hex_to_sha(hex)
+ except (TypeError, AssertionError):
+ raise ObjectFormatException("%s %s" % (error_msg, hex))
+
+
+def check_identity(identity, error_msg):
+ email_start = identity.find("<")
+ email_end = identity.find(">")
+ if (email_start < 0 or email_end < 0 or email_end <= email_start
+ or identity.find("<", email_start + 1) >= 0
+ or identity.find(">", email_end + 1) >= 0
+ or not identity.endswith(">")):
+ raise ObjectFormatException(error_msg)
+
+
+class FixedSha(object):
+ """SHA object that behaves like hashlib's but is given a fixed value."""
+
+ def __init__(self, hexsha):
+ self._hexsha = hexsha
+ self._sha = hex_to_sha(hexsha)
+
+ def digest(self):
+ return self._sha
+
+ def hexdigest(self):
+ return self._hexsha
class ShaFile(object):
"""A git SHA file."""
-
- @classmethod
- def _parse_legacy_object(cls, map):
- """Parse a legacy object, creating it and setting object._text"""
- text = _decompress(map)
- object = None
- for posstype in type_map.keys():
- if text.startswith(posstype):
- object = type_map[posstype]()
- text = text[len(posstype):]
- break
- assert object is not None, "%s is not a known object type" % text[:9]
- assert text[0] == ' ', "%s is not a space" % text[0]
- text = text[1:]
- size = 0
- i = 0
- while text[0] >= '0' and text[0] <= '9':
- if i > 0 and size == 0:
- assert False, "Size is not in canonical format"
- size = (size * 10) + int(text[0])
- text = text[1:]
- i += 1
- object._size = size
- assert text[0] == "\0", "Size not followed by null"
- text = text[1:]
- object._text = text
- return object
+
+ @staticmethod
+ def _parse_legacy_object_header(magic, f):
+ """Parse a legacy object, creating it but not reading the file."""
+ bufsize = 1024
+ decomp = zlib.decompressobj()
+ header = decomp.decompress(magic)
+ start = 0
+ end = -1
+ while end < 0:
+ header += decomp.decompress(f.read(bufsize))
+ end = header.find("\0", start)
+ start = len(header)
+ header = header[:end]
+ type_name, size = header.split(" ", 1)
+ size = int(size) # sanity check
+ obj_class = object_class(type_name)
+ if not obj_class:
+ raise ObjectFormatException("Not a known type: %s" % type_name)
+ obj = obj_class()
+ obj._filename = f.name
+ return obj
+
+ def _parse_legacy_object(self, f):
+ """Parse a legacy object, setting the raw string."""
+ size = os.path.getsize(f.name)
+ map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
+ try:
+ text = _decompress(map)
+ finally:
+ map.close()
+ header_end = text.find('\0')
+ if header_end < 0:
+ raise ObjectFormatException("Invalid object header")
+ self.set_raw_string(text[header_end+1:])
+
+ def as_legacy_object_chunks(self):
+ compobj = zlib.compressobj()
+ yield compobj.compress(self._header())
+ for chunk in self.as_raw_chunks():
+ yield compobj.compress(chunk)
+ yield compobj.flush()
def as_legacy_object(self):
- return zlib.compress("%s %d\0%s" % (self._type, len(self._text), self._text))
-
+ return "".join(self.as_legacy_object_chunks())
+
+ def as_raw_chunks(self):
+ if self._needs_parsing:
+ self._ensure_parsed()
+ elif self._needs_serialization:
+ self._chunked_text = self._serialize()
+ return self._chunked_text
+
def as_raw_string(self):
- return self._num_type, self._text
-
- @classmethod
- def _parse_object(cls, map):
- """Parse a new style object , creating it and setting object._text"""
- used = 0
- byte = ord(map[used])
- used += 1
- num_type = (byte >> 4) & 7
+ return "".join(self.as_raw_chunks())
+
+ def __str__(self):
+ return self.as_raw_string()
+
+ def __hash__(self):
+ return hash(self.id)
+
+ def as_pretty_string(self):
+ return self.as_raw_string()
+
+ def _ensure_parsed(self):
+ if self._needs_parsing:
+ if not self._chunked_text:
+ assert self._filename, "ShaFile needs either text or filename"
+ self._parse_file()
+ self._deserialize(self._chunked_text)
+ self._needs_parsing = False
+
+ def set_raw_string(self, text):
+ if type(text) != str:
+ raise TypeError(text)
+ self.set_raw_chunks([text])
+
+ def set_raw_chunks(self, chunks):
+ self._chunked_text = chunks
+ self._deserialize(chunks)
+ self._sha = None
+ self._needs_parsing = False
+ self._needs_serialization = False
+
+ @staticmethod
+ def _parse_object_header(magic, f):
+ """Parse a new style object, creating it but not reading the file."""
+ num_type = (ord(magic[0]) >> 4) & 7
+ obj_class = object_class(num_type)
+ if not obj_class:
+ raise ObjectFormatException("Not a known type: %d" % num_type)
+ obj = obj_class()
+ obj._filename = f.name
+ return obj
+
+ def _parse_object(self, f):
+ """Parse a new style object, setting self._text."""
+ size = os.path.getsize(f.name)
+ map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
try:
- object = num_type_map[num_type]()
- except KeyError:
- raise AssertionError("Not a known type: %d" % num_type)
- while (byte & 0x80) != 0:
- byte = ord(map[used])
- used += 1
- raw = map[used:]
- object._text = _decompress(raw)
- return object
-
+ # skip type and size; type must have already been determined, and we
+ # trust zlib to fail if it's otherwise corrupted
+ byte = ord(map[0])
+ used = 1
+ while (byte & 0x80) != 0:
+ byte = ord(map[used])
+ used += 1
+ raw = map[used:]
+ self.set_raw_string(_decompress(raw))
+ finally:
+ map.close()
+
+ @classmethod
+ def _is_legacy_object(cls, magic):
+ b0, b1 = map(ord, magic)
+ word = (b0 << 8) + b1
+ return b0 == 0x78 and (word % 31) == 0
+
@classmethod
- def _parse_file(cls, map):
- word = (ord(map[0]) << 8) + ord(map[1])
- if ord(map[0]) == 0x78 and (word % 31) == 0:
- return cls._parse_legacy_object(map)
+ def _parse_file_header(cls, f):
+ magic = f.read(2)
+ if cls._is_legacy_object(magic):
+ return cls._parse_legacy_object_header(magic, f)
else:
- return cls._parse_object(map)
-
+ return cls._parse_object_header(magic, f)
+
def __init__(self):
"""Don't call this directly"""
-
- def _parse_text(self):
- """For subclasses to do initialisation time parsing"""
-
+ self._sha = None
+ self._filename = None
+ self._chunked_text = []
+ self._needs_parsing = False
+ self._needs_serialization = True
+
+ def _deserialize(self, chunks):
+ raise NotImplementedError(self._deserialize)
+
+ def _serialize(self):
+ raise NotImplementedError(self._serialize)
+
+ def _parse_file(self):
+ f = GitFile(self._filename, 'rb')
+ try:
+ magic = f.read(2)
+ if self._is_legacy_object(magic):
+ self._parse_legacy_object(f)
+ else:
+ self._parse_object(f)
+ finally:
+ f.close()
+
@classmethod
- def from_file(cls, filename):
- """Get the contents of a SHA file on disk"""
- size = os.path.getsize(filename)
- f = open(filename, 'rb')
+ def from_path(cls, path):
+ f = GitFile(path, 'rb')
try:
- map = mmap.mmap(f.fileno(), size, access=mmap.ACCESS_READ)
- shafile = cls._parse_file(map)
- shafile._parse_text()
- return shafile
+ obj = cls.from_file(f)
+ obj._sha = FixedSha(filename_to_hex(path))
+ return obj
finally:
f.close()
-
+
@classmethod
- def from_raw_string(cls, type, string):
+ def from_file(cls, f):
+ """Get the contents of a SHA file on disk."""
+ try:
+ obj = cls._parse_file_header(f)
+ obj._sha = None
+ obj._needs_parsing = True
+ obj._needs_serialization = True
+ return obj
+ except (IndexError, ValueError), e:
+ raise ObjectFormatException("invalid object header")
+
+ @staticmethod
+ def from_raw_string(type_num, string):
"""Creates an object of the indicated type from the raw string given.
-
- Type is the numeric type of an object. String is the raw uncompressed
- contents.
+
+ :param type_num: The numeric type of the object.
+ :param string: The raw uncompressed contents.
+ """
+ obj = object_class(type_num)()
+ obj.set_raw_string(string)
+ return obj
+
+ @staticmethod
+ def from_raw_chunks(type_num, chunks):
+ """Creates an object of the indicated type from the raw chunks given.
+
+ :param type_num: The numeric type of the object.
+ :param chunks: An iterable of the raw uncompressed contents.
"""
- real_class = num_type_map[type]
- obj = real_class()
- obj._num_type = type
- obj._text = string
- obj._parse_text()
+ obj = object_class(type_num)()
+ obj.set_raw_chunks(chunks)
+ return obj
+
+ @classmethod
+ def from_string(cls, string):
+ """Create a ShaFile from a string."""
+ obj = cls()
+ obj.set_raw_string(string)
return obj
-
+
+ def _check_has_member(self, member, error_msg):
+ """Check that the object has a given member variable.
+
+ :param member: the member variable to check for
+ :param error_msg: the message for an error if the member is missing
+ :raise ObjectFormatException: with the given error_msg if member is
+ missing or is None
+ """
+ if getattr(self, member, None) is None:
+ raise ObjectFormatException(error_msg)
+
+ def check(self):
+ """Check this object for internal consistency.
+
+ :raise ObjectFormatException: if the object is malformed in some way
+ :raise ChecksumMismatch: if the object was created with a SHA that does
+ not match its contents
+ """
+ # TODO: if we find that error-checking during object parsing is a
+ # performance bottleneck, those checks should be moved to the class's
+ # check() method during optimization so we can still check the object
+ # when necessary.
+ old_sha = self.id
+ try:
+ self._deserialize(self.as_raw_chunks())
+ self._sha = None
+ new_sha = self.id
+ except Exception, e:
+ raise ObjectFormatException(e)
+ if old_sha != new_sha:
+ raise ChecksumMismatch(new_sha, old_sha)
+
def _header(self):
- return "%s %lu\0" % (self._type, len(self._text))
-
+ return object_header(self.type, self.raw_length())
+
+ def raw_length(self):
+ """Returns the length of the raw string of this object."""
+ ret = 0
+ for chunk in self.as_raw_chunks():
+ ret += len(chunk)
+ return ret
+
+ def _make_sha(self):
+ ret = make_sha()
+ ret.update(self._header())
+ for chunk in self.as_raw_chunks():
+ ret.update(chunk)
+ return ret
+
def sha(self):
"""The SHA1 object that is the name of this object."""
- ressha = sha.new()
- ressha.update(self._header())
- ressha.update(self._text)
- return ressha
-
+ if self._sha is None:
+ # this is a local because as_raw_chunks() overwrites self._sha
+ new_sha = make_sha()
+ new_sha.update(self._header())
+ for chunk in self.as_raw_chunks():
+ new_sha.update(chunk)
+ self._sha = new_sha
+ return self._sha
+
@property
def id(self):
return self.sha().hexdigest()
-
- @property
- def type(self):
- return self._num_type
-
+
+ def get_type(self):
+ return self.type_num
+
+ def set_type(self, type):
+ self.type_num = type
+
+ # DEPRECATED: use type_num or type_name as needed.
+ type = property(get_type, set_type)
+
def __repr__(self):
return "<%s %s>" % (self.__class__.__name__, self.id)
-
+
+ def __ne__(self, other):
+ return self.id != other.id
+
def __eq__(self, other):
- """Return true id the sha of the two objects match.
-
+ """Return true if the sha of the two objects match.
+
The __le__ etc methods aren't overriden as they make no sense,
certainly at this level.
"""
- return self.sha().digest() == other.sha().digest()
+ return self.id == other.id
class Blob(ShaFile):
"""A Git Blob object."""
- _type = BLOB_ID
- _num_type = 3
+ type_name = 'blob'
+ type_num = 3
- @property
- def data(self):
- """The text contained within the blob object."""
- return self._text
+ def __init__(self):
+ super(Blob, self).__init__()
+ self._chunked_text = []
+ self._needs_parsing = False
+ self._needs_serialization = False
+
+ def _get_data(self):
+ return self.as_raw_string()
+
+ def _set_data(self, data):
+ self.set_raw_string(data)
+
+ data = property(_get_data, _set_data,
+ "The text contained within the blob object.")
+
+ def _get_chunked(self):
+ self._ensure_parsed()
+ return self._chunked_text
+
+ def _set_chunked(self, chunks):
+ self._chunked_text = chunks
+
+ def _serialize(self):
+ if not self._chunked_text:
+ self._ensure_parsed()
+ self._needs_serialization = False
+ return self._chunked_text
+
+ def _deserialize(self, chunks):
+ self._chunked_text = chunks
+
+ chunked = property(_get_chunked, _set_chunked,
+ "The text within the blob object, as chunks (not necessarily lines).")
@classmethod
- def from_file(cls, filename):
- blob = ShaFile.from_file(filename)
- if blob._type != cls._type:
- raise NotBlobError(filename)
+ def from_path(cls, path):
+ blob = ShaFile.from_path(path)
+ if not isinstance(blob, cls):
+ raise NotBlobError(path)
return blob
- @classmethod
- def from_string(cls, string):
- """Create a blob from a string."""
- shafile = cls()
- shafile._text = string
- return shafile
+ def check(self):
+ """Check this object for internal consistency.
+
+ :raise ObjectFormatException: if the object is malformed in some way
+ """
+ super(Blob, self).check()
+
+
+def _parse_tag_or_commit(text):
+ """Parse tag or commit text.
+
+ :param text: the raw text of the tag or commit object.
+ :yield: tuples of (field, value), one per header line, in the order read
+ from the text, possibly including duplicates. Includes a field named
+ None for the freeform tag/commit text.
+ """
+ f = StringIO(text)
+ for l in f:
+ l = l.rstrip("\n")
+ if l == "":
+ # Empty line indicates end of headers
+ break
+ yield l.split(" ", 1)
+ yield (None, f.read())
+ f.close()
+
+
+def parse_tag(text):
+ return _parse_tag_or_commit(text)
class Tag(ShaFile):
"""A Git Tag object."""
- _type = TAG_ID
- _num_type = 4
+ type_name = 'tag'
+ type_num = 4
- @classmethod
- def from_file(cls, filename):
- blob = ShaFile.from_file(filename)
- if blob._type != cls._type:
- raise NotBlobError(filename)
- return blob
+ def __init__(self):
+ super(Tag, self).__init__()
+ self._tag_timezone_neg_utc = False
@classmethod
- def from_string(cls, string):
- """Create a blob from a string."""
- shafile = cls()
- shafile._text = string
- return shafile
+ def from_path(cls, filename):
+ tag = ShaFile.from_path(filename)
+ if not isinstance(tag, cls):
+ raise NotTagError(filename)
+ return tag
+
+ def check(self):
+ """Check this object for internal consistency.
- def _parse_text(self):
+ :raise ObjectFormatException: if the object is malformed in some way
+ """
+ super(Tag, self).check()
+ self._check_has_member("_object_sha", "missing object sha")
+ self._check_has_member("_object_class", "missing object type")
+ self._check_has_member("_name", "missing tag name")
+
+ if not self._name:
+ raise ObjectFormatException("empty tag name")
+
+ check_hexsha(self._object_sha, "invalid object sha")
+
+ if getattr(self, "_tagger", None):
+ check_identity(self._tagger, "invalid tagger")
+
+ last = None
+ for field, _ in parse_tag("".join(self._chunked_text)):
+ if field == _OBJECT_HEADER and last is not None:
+ raise ObjectFormatException("unexpected object")
+ elif field == _TYPE_HEADER and last != _OBJECT_HEADER:
+ raise ObjectFormatException("unexpected type")
+ elif field == _TAG_HEADER and last != _TYPE_HEADER:
+ raise ObjectFormatException("unexpected tag name")
+ elif field == _TAGGER_HEADER and last != _TAG_HEADER:
+ raise ObjectFormatException("unexpected tagger")
+ last = field
+
+ def _serialize(self):
+ chunks = []
+ chunks.append("%s %s\n" % (_OBJECT_HEADER, self._object_sha))
+ chunks.append("%s %s\n" % (_TYPE_HEADER, self._object_class.type_name))
+ chunks.append("%s %s\n" % (_TAG_HEADER, self._name))
+ if self._tagger:
+ if self._tag_time is None:
+ chunks.append("%s %s\n" % (_TAGGER_HEADER, self._tagger))
+ else:
+ chunks.append("%s %s %d %s\n" % (
+ _TAGGER_HEADER, self._tagger, self._tag_time,
+ format_timezone(self._tag_timezone,
+ self._tag_timezone_neg_utc)))
+ chunks.append("\n") # To close headers
+ chunks.append(self._message)
+ return chunks
+
+ def _deserialize(self, chunks):
"""Grab the metadata attached to the tag"""
- text = self._text
- count = 0
- assert text.startswith(OBJECT_ID), "Invalid tag object, " \
- "must start with %s" % OBJECT_ID
- count += len(OBJECT_ID)
- assert text[count] == ' ', "Invalid tag object, " \
- "%s must be followed by space not %s" % (OBJECT_ID, text[count])
- count += 1
- self._object_sha = text[count:count+40]
- count += 40
- assert text[count] == '\n', "Invalid tag object, " \
- "%s sha must be followed by newline" % OBJECT_ID
- count += 1
- assert text[count:].startswith(TYPE_ID), "Invalid tag object, " \
- "%s sha must be followed by %s" % (OBJECT_ID, TYPE_ID)
- count += len(TYPE_ID)
- assert text[count] == ' ', "Invalid tag object, " \
- "%s must be followed by space not %s" % (TAG_ID, text[count])
- count += 1
- self._object_type = ""
- while text[count] != '\n':
- self._object_type += text[count]
- count += 1
- count += 1
- assert self._object_type in (COMMIT_ID, BLOB_ID, TREE_ID, TAG_ID), "Invalid tag object, " \
- "unexpected object type %s" % self._object_type
- self._object_type = type_map[self._object_type]
-
- assert text[count:].startswith(TAG_ID), "Invalid tag object, " \
- "object type must be followed by %s" % (TAG_ID)
- count += len(TAG_ID)
- assert text[count] == ' ', "Invalid tag object, " \
- "%s must be followed by space not %s" % (TAG_ID, text[count])
- count += 1
- self._name = ""
- while text[count] != '\n':
- self._name += text[count]
- count += 1
- count += 1
-
- assert text[count:].startswith(TAGGER_ID), "Invalid tag object, " \
- "%s must be followed by %s" % (TAG_ID, TAGGER_ID)
- count += len(TAGGER_ID)
- assert text[count] == ' ', "Invalid tag object, " \
- "%s must be followed by space not %s" % (TAGGER_ID, text[count])
- count += 1
- self._tagger = ""
- while text[count] != '>':
- assert text[count] != '\n', "Malformed tagger information"
- self._tagger += text[count]
- count += 1
- self._tagger += text[count]
- count += 1
- assert text[count] == ' ', "Invalid tag object, " \
- "tagger information must be followed by space not %s" % text[count]
- count += 1
- self._tag_time = int(text[count:count+10])
- while text[count] != '\n':
- count += 1
- count += 1
- assert text[count] == '\n', "There must be a new line after the headers"
- count += 1
- self._message = text[count:]
+ self._tagger = None
+ for field, value in parse_tag("".join(chunks)):
+ if field == _OBJECT_HEADER:
+ self._object_sha = value
+ elif field == _TYPE_HEADER:
+ obj_class = object_class(value)
+ if not obj_class:
+ raise ObjectFormatException("Not a known type: %s" % value)
+ self._object_class = obj_class
+ elif field == _TAG_HEADER:
+ self._name = value
+ elif field == _TAGGER_HEADER:
+ try:
+ sep = value.index("> ")
+ except ValueError:
+ self._tagger = value
+ self._tag_time = None
+ self._tag_timezone = None
+ self._tag_timezone_neg_utc = False
+ else:
+ self._tagger = value[0:sep+1]
+ try:
+ (timetext, timezonetext) = value[sep+2:].rsplit(" ", 1)
+ self._tag_time = int(timetext)
+ self._tag_timezone, self._tag_timezone_neg_utc = \
+ parse_timezone(timezonetext)
+ except ValueError, e:
+ raise ObjectFormatException(e)
+ elif field is None:
+ self._message = value
+ else:
+ raise ObjectFormatException("Unknown field %s" % field)
+
+ def _get_object(self):
+ """Get the object pointed to by this tag.
+
+ :return: tuple of (object class, sha).
+ """
+ self._ensure_parsed()
+ return (self._object_class, self._object_sha)
- @property
- def object(self):
- """Returns the object pointed by this tag, represented as a tuple(type, sha)"""
- return (self._object_type, self._object_sha)
+ def _set_object(self, value):
+ self._ensure_parsed()
+ (self._object_class, self._object_sha) = value
+ self._needs_serialization = True
- @property
- def name(self):
- """Returns the name of this tag"""
- return self._name
+ object = property(_get_object, _set_object)
- @property
- def tagger(self):
- """Returns the name of the person who created this tag"""
- return self._tagger
+ name = serializable_property("name", "The name of this tag")
+ tagger = serializable_property("tagger",
+ "Returns the name of the person who created this tag")
+ tag_time = serializable_property("tag_time",
+ "The creation timestamp of the tag. As the number of seconds since the epoch")
+ tag_timezone = serializable_property("tag_timezone",
+ "The timezone that tag_time is in.")
+ message = serializable_property("message", "The message attached to this tag")
- @property
- def tag_time(self):
- """Returns the creation timestamp of the tag.
- Returns it as the number of seconds since the epoch"""
- return self._tag_time
+def parse_tree(text):
+ """Parse a tree text.
+
+ :param text: Serialized text to parse
+ :yields: tuples of (name, mode, sha)
+ """
+ count = 0
+ l = len(text)
+ while count < l:
+ mode_end = text.index(' ', count)
+ mode = int(text[count:mode_end], 8)
+ name_end = text.index('\0', mode_end)
+ name = text[mode_end+1:name_end]
+ count = name_end+21
+ sha = text[name_end+1:count]
+ yield (name, mode, sha_to_hex(sha))
- @property
- def message(self):
- """Returns the message attached to this tag"""
- return self._message
+def serialize_tree(items):
+ """Serialize the items in a tree to a text.
-def parse_tree(text):
- ret = []
- count = 0
- while count < len(text):
- mode = 0
- chr = text[count]
- while chr != ' ':
- assert chr >= '0' and chr <= '7', "%s is not a valid mode char" % chr
- mode = (mode << 3) + (ord(chr) - ord('0'))
- count += 1
- chr = text[count]
- count += 1
- chr = text[count]
- name = ''
- while chr != '\0':
- name += chr
- count += 1
- chr = text[count]
- count += 1
- chr = text[count]
- sha = text[count:count+20]
- hexsha = sha_to_hex(sha)
- ret.append((mode, name, hexsha))
- count = count + 20
- return ret
+ :param items: Sorted iterable over (name, mode, sha) tuples
+ :return: Serialized tree text as chunks
+ """
+ for name, mode, hexsha in items:
+ yield "%04o %s\0%s" % (mode, name, hex_to_sha(hexsha))
+
+
+def sorted_tree_items(entries):
+ """Iterate over a tree entries dictionary in the order in which
+ the items would be serialized.
+
+ :param entries: Dictionary mapping names to (mode, sha) tuples
+ :return: Iterator over (name, mode, sha)
+ """
+ for name, entry in sorted(entries.iteritems(), cmp=cmp_entry):
+ yield name, entry[0], entry[1]
+
+
+def cmp_entry((name1, value1), (name2, value2)):
+ """Compare two tree entries."""
+ if stat.S_ISDIR(value1[0]):
+ name1 += "/"
+ if stat.S_ISDIR(value2[0]):
+ name2 += "/"
+ return cmp(name1, name2)
class Tree(ShaFile):
"""A Git tree object"""
- _type = TREE_ID
- _num_type = 2
+ type_name = 'tree'
+ type_num = 2
def __init__(self):
+ super(Tree, self).__init__()
self._entries = {}
@classmethod
- def from_file(cls, filename):
- tree = ShaFile.from_file(filename)
- if tree._type != cls._type:
+ def from_path(cls, filename):
+ tree = ShaFile.from_path(filename)
+ if not isinstance(tree, cls):
raise NotTreeError(filename)
return tree
+ def __contains__(self, name):
+ self._ensure_parsed()
+ return name in self._entries
+
def __getitem__(self, name):
+ self._ensure_parsed()
return self._entries[name]
def __setitem__(self, name, value):
assert isinstance(value, tuple)
assert len(value) == 2
+ self._ensure_parsed()
self._entries[name] = value
+ self._needs_serialization = True
def __delitem__(self, name):
+ self._ensure_parsed()
del self._entries[name]
+ self._needs_serialization = True
+
+ def __len__(self):
+ self._ensure_parsed()
+ return len(self._entries)
+
+ def __iter__(self):
+ self._ensure_parsed()
+ return iter(self._entries)
def add(self, mode, name, hexsha):
+ assert type(mode) == int
+ assert type(name) == str
+ assert type(hexsha) == str
+ self._ensure_parsed()
self._entries[name] = mode, hexsha
+ self._needs_serialization = True
def entries(self):
"""Return a list of tuples describing the tree entries"""
- # The order of this is different from iteritems() for historical reasons
- return [(mode, name, hexsha) for (name, mode, hexsha) in self.iteritems()]
+ self._ensure_parsed()
+ # The order of this is different from iteritems() for historical
+ # reasons
+ return [
+ (mode, name, hexsha) for (name, mode, hexsha) in self.iteritems()]
def iteritems(self):
- for name in sorted(self._entries.keys()):
- yield name, self._entries[name][0], self._entries[name][1]
+ """Iterate over entries in the order in which they would be serialized.
+
+ :return: Iterator over (name, mode, sha) tuples
+ """
+ self._ensure_parsed()
+ return sorted_tree_items(self._entries)
- def _parse_text(self):
+ def _deserialize(self, chunks):
"""Grab the entries in the tree"""
- self._entries = parse_tree(self._text)
+ try:
+ parsed_entries = parse_tree("".join(chunks))
+ except ValueError, e:
+ raise ObjectFormatException(e)
+ # TODO: list comprehension is for efficiency in the common (small) case;
+ # if memory efficiency in the large case is a concern, use a genexp.
+ self._entries = dict([(n, (m, s)) for n, m, s in parsed_entries])
+
+ def check(self):
+ """Check this object for internal consistency.
- def serialize(self):
- self._text = ""
+ :raise ObjectFormatException: if the object is malformed in some way
+ """
+ super(Tree, self).check()
+ last = None
+ allowed_modes = (stat.S_IFREG | 0755, stat.S_IFREG | 0644,
+ stat.S_IFLNK, stat.S_IFDIR, S_IFGITLINK,
+ # TODO: optionally exclude as in git fsck --strict
+ stat.S_IFREG | 0664)
+ for name, mode, sha in parse_tree("".join(self._chunked_text)):
+ check_hexsha(sha, 'invalid sha %s' % sha)
+ if '/' in name or name in ('', '.', '..'):
+ raise ObjectFormatException('invalid name %s' % name)
+
+ if mode not in allowed_modes:
+ raise ObjectFormatException('invalid mode %06o' % mode)
+
+ entry = (name, (mode, sha))
+ if last:
+ if cmp_entry(last, entry) > 0:
+ raise ObjectFormatException('entries not sorted')
+ if name == last[0]:
+ raise ObjectFormatException('duplicate entry %s' % name)
+ last = entry
+
+ def _serialize(self):
+ return list(serialize_tree(self.iteritems()))
+
+ def as_pretty_string(self):
+ text = []
for name, mode, hexsha in self.iteritems():
- self._text += "%04o %s\0%s" % (mode, name, hex_to_sha(hexsha))
+ if mode & stat.S_IFDIR:
+ kind = "tree"
+ else:
+ kind = "blob"
+ text.append("%04o %s %s\t%s\n" % (mode, kind, hexsha, name))
+ return "".join(text)
+
+
+def parse_timezone(text):
+ offset = int(text)
+ negative_utc = (offset == 0 and text[0] == '-')
+ signum = (offset < 0) and -1 or 1
+ offset = abs(offset)
+ hours = int(offset / 100)
+ minutes = (offset % 100)
+ return signum * (hours * 3600 + minutes * 60), negative_utc
+
+
+def format_timezone(offset, negative_utc=False):
+ if offset % 60 != 0:
+ raise ValueError("Unable to handle non-minute offset.")
+ if offset < 0 or (offset == 0 and negative_utc):
+ sign = '-'
+ else:
+ sign = '+'
+ offset = abs(offset)
+ return '%c%02d%02d' % (sign, offset / 3600, (offset / 60) % 60)
+
+
+def parse_commit(text):
+ return _parse_tag_or_commit(text)
class Commit(ShaFile):
"""A git commit object"""
- _type = COMMIT_ID
- _num_type = 1
+ type_name = 'commit'
+ type_num = 1
def __init__(self):
+ super(Commit, self).__init__()
self._parents = []
+ self._encoding = None
+ self._extra = {}
+ self._author_timezone_neg_utc = False
+ self._commit_timezone_neg_utc = False
@classmethod
- def from_file(cls, filename):
- commit = ShaFile.from_file(filename)
- if commit._type != cls._type:
- raise NotCommitError(filename)
+ def from_path(cls, path):
+ commit = ShaFile.from_path(path)
+ if not isinstance(commit, cls):
+ raise NotCommitError(path)
return commit
- def _parse_text(self):
- text = self._text
- count = 0
- assert text.startswith(TREE_ID), "Invalid commit object, " \
- "must start with %s" % TREE_ID
- count += len(TREE_ID)
- assert text[count] == ' ', "Invalid commit object, " \
- "%s must be followed by space not %s" % (TREE_ID, text[count])
- count += 1
- self._tree = text[count:count+40]
- count = count + 40
- assert text[count] == "\n", "Invalid commit object, " \
- "tree sha must be followed by newline"
- count += 1
+ def _deserialize(self, chunks):
self._parents = []
- while text[count:].startswith(PARENT_ID):
- count += len(PARENT_ID)
- assert text[count] == ' ', "Invalid commit object, " \
- "%s must be followed by space not %s" % (PARENT_ID, text[count])
- count += 1
- self._parents.append(text[count:count+40])
- count += 40
- assert text[count] == "\n", "Invalid commit object, " \
- "parent sha must be followed by newline"
- count += 1
+ self._extra = []
self._author = None
- if text[count:].startswith(AUTHOR_ID):
- count += len(AUTHOR_ID)
- assert text[count] == ' ', "Invalid commit object, " \
- "%s must be followed by space not %s" % (AUTHOR_ID, text[count])
- count += 1
- self._author = ''
- while text[count] != '>':
- assert text[count] != '\n', "Malformed author information"
- self._author += text[count]
- count += 1
- self._author += text[count]
- count += 1
- assert text[count] == ' ', "Invalid commit object, " \
- "author information must be followed by space not %s" % text[count]
- count += 1
- self._author_time = int(text[count:count+10])
- while text[count] != ' ':
- assert text[count] != '\n', "Malformed author information"
- count += 1
- self._author_timezone = int(text[count:count+6])
- count += 1
- while text[count] != '\n':
- count += 1
- count += 1
- self._committer = None
- if text[count:].startswith(COMMITTER_ID):
- count += len(COMMITTER_ID)
- assert text[count] == ' ', "Invalid commit object, " \
- "%s must be followed by space not %s" % (COMMITTER_ID, text[count])
- count += 1
- self._committer = ''
- while text[count] != '>':
- assert text[count] != '\n', "Malformed committer information"
- self._committer += text[count]
- count += 1
- self._committer += text[count]
- count += 1
- assert text[count] == ' ', "Invalid commit object, " \
- "commiter information must be followed by space not %s" % text[count]
- count += 1
- self._commit_time = int(text[count:count+10])
- while text[count] != ' ':
- assert text[count] != '\n', "Malformed committer information"
- count += 1
- self._commit_timezone = int(text[count:count+6])
- count += 1
- while text[count] != '\n':
- count += 1
- count += 1
- assert text[count] == '\n', "There must be a new line after the headers"
- count += 1
- # XXX: There can be an encoding field.
- self._message = text[count:]
-
- def serialize(self):
- self._text = ""
- self._text += "%s %s\n" % (TREE_ID, self._tree)
+ for field, value in parse_commit("".join(self._chunked_text)):
+ if field == _TREE_HEADER:
+ self._tree = value
+ elif field == _PARENT_HEADER:
+ self._parents.append(value)
+ elif field == _AUTHOR_HEADER:
+ self._author, timetext, timezonetext = value.rsplit(" ", 2)
+ self._author_time = int(timetext)
+ self._author_timezone, self._author_timezone_neg_utc =\
+ parse_timezone(timezonetext)
+ elif field == _COMMITTER_HEADER:
+ self._committer, timetext, timezonetext = value.rsplit(" ", 2)
+ self._commit_time = int(timetext)
+ self._commit_timezone, self._commit_timezone_neg_utc =\
+ parse_timezone(timezonetext)
+ elif field == _ENCODING_HEADER:
+ self._encoding = value
+ elif field is None:
+ self._message = value
+ else:
+ self._extra.append((field, value))
+
+ def check(self):
+ """Check this object for internal consistency.
+
+ :raise ObjectFormatException: if the object is malformed in some way
+ """
+ super(Commit, self).check()
+ self._check_has_member("_tree", "missing tree")
+ self._check_has_member("_author", "missing author")
+ self._check_has_member("_committer", "missing committer")
+ # times are currently checked when set
+
+ for parent in self._parents:
+ check_hexsha(parent, "invalid parent sha")
+ check_hexsha(self._tree, "invalid tree sha")
+
+ check_identity(self._author, "invalid author")
+ check_identity(self._committer, "invalid committer")
+
+ last = None
+ for field, _ in parse_commit("".join(self._chunked_text)):
+ if field == _TREE_HEADER and last is not None:
+ raise ObjectFormatException("unexpected tree")
+ elif field == _PARENT_HEADER and last not in (_PARENT_HEADER,
+ _TREE_HEADER):
+ raise ObjectFormatException("unexpected parent")
+ elif field == _AUTHOR_HEADER and last not in (_TREE_HEADER,
+ _PARENT_HEADER):
+ raise ObjectFormatException("unexpected author")
+ elif field == _COMMITTER_HEADER and last != _AUTHOR_HEADER:
+ raise ObjectFormatException("unexpected committer")
+ elif field == _ENCODING_HEADER and last != _COMMITTER_HEADER:
+ raise ObjectFormatException("unexpected encoding")
+ last = field
+
+ # TODO: optionally check for duplicate parents
+
+ def _serialize(self):
+ chunks = []
+ chunks.append("%s %s\n" % (_TREE_HEADER, self._tree))
for p in self._parents:
- self._text += "%s %s\n" % (PARENT_ID, p)
- self._text += "%s %s %s %+05d\n" % (AUTHOR_ID, self._author, str(self._author_time), self._author_timezone)
- self._text += "%s %s %s %+05d\n" % (COMMITTER_ID, self._committer, str(self._commit_time), self._commit_timezone)
- self._text += "\n" # There must be a new line after the headers
- self._text += self._message
-
- @property
- def tree(self):
- """Returns the tree that is the state of this commit"""
- return self._tree
-
- @property
- def parents(self):
+ chunks.append("%s %s\n" % (_PARENT_HEADER, p))
+ chunks.append("%s %s %s %s\n" % (
+ _AUTHOR_HEADER, self._author, str(self._author_time),
+ format_timezone(self._author_timezone,
+ self._author_timezone_neg_utc)))
+ chunks.append("%s %s %s %s\n" % (
+ _COMMITTER_HEADER, self._committer, str(self._commit_time),
+ format_timezone(self._commit_timezone,
+ self._commit_timezone_neg_utc)))
+ if self.encoding:
+ chunks.append("%s %s\n" % (_ENCODING_HEADER, self.encoding))
+ for k, v in self.extra:
+ if "\n" in k or "\n" in v:
+ raise AssertionError("newline in extra data: %r -> %r" % (k, v))
+ chunks.append("%s %s\n" % (k, v))
+ chunks.append("\n") # There must be a new line after the headers
+ chunks.append(self._message)
+ return chunks
+
+ tree = serializable_property("tree", "Tree that is the state of this commit")
+
+ def _get_parents(self):
"""Return a list of parents of this commit."""
+ self._ensure_parsed()
return self._parents
- @property
- def author(self):
- """Returns the name of the author of the commit"""
- return self._author
+ def _set_parents(self, value):
+ """Set a list of parents of this commit."""
+ self._ensure_parsed()
+ self._needs_serialization = True
+ self._parents = value
- @property
- def committer(self):
- """Returns the name of the committer of the commit"""
- return self._committer
+ parents = property(_get_parents, _set_parents)
- @property
- def message(self):
- """Returns the commit message"""
- return self._message
+ def _get_extra(self):
+ """Return extra settings of this commit."""
+ self._ensure_parsed()
+ return self._extra
- @property
- def commit_time(self):
- """Returns the timestamp of the commit.
-
- Returns it as the number of seconds since the epoch.
- """
- return self._commit_time
+ extra = property(_get_extra)
- @property
- def commit_timezone(self):
- """Returns the zone the commit time is in
- """
- return self._commit_timezone
+ author = serializable_property("author",
+ "The name of the author of the commit")
- @property
- def author_time(self):
- """Returns the timestamp the commit was written.
-
- Returns it as the number of seconds since the epoch.
- """
- return self._author_time
+ committer = serializable_property("committer",
+ "The name of the committer of the commit")
- @property
- def author_timezone(self):
- """Returns the zone the author time is in
- """
- return self._author_timezone
+ message = serializable_property("message",
+ "The commit message")
+ commit_time = serializable_property("commit_time",
+ "The timestamp of the commit. As the number of seconds since the epoch.")
-type_map = {
- BLOB_ID : Blob,
- TREE_ID : Tree,
- COMMIT_ID : Commit,
- TAG_ID: Tag,
-}
+ commit_timezone = serializable_property("commit_timezone",
+ "The zone the commit time is in")
-num_type_map = {
- 0: None,
- 1: Commit,
- 2: Tree,
- 3: Blob,
- 4: Tag,
- # 5 Is reserved for further expansion
-}
+ author_time = serializable_property("author_time",
+ "The timestamp the commit was written. as the number of seconds since the epoch.")
+ author_timezone = serializable_property("author_timezone",
+ "Returns the zone the author time is in.")
+
+ encoding = serializable_property("encoding",
+ "Encoding of the commit message.")
+
+
+OBJECT_CLASSES = (
+ Commit,
+ Tree,
+ Blob,
+ Tag,
+ )
+
+_TYPE_MAP = {}
+
+for cls in OBJECT_CLASSES:
+ _TYPE_MAP[cls.type_name] = cls
+ _TYPE_MAP[cls.type_num] = cls
+
+
+
+# Hold on to the pure-python implementations for testing
+_parse_tree_py = parse_tree
+_sorted_tree_items_py = sorted_tree_items
try:
# Try to import C versions
- from dulwich._objects import hex_to_sha, sha_to_hex, parse_tree
+ from dulwich._objects import parse_tree, sorted_tree_items
except ImportError:
pass
-