/usr/lib/python3/dist-packages/gitdb/pack.py is in python3-gitdb 2.0.3-1.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 | # Copyright (C) 2010, 2011 Sebastian Thiel (byronimo@gmail.com) and contributors
#
# This module is part of GitDB and is released under
# the New BSD License: http://www.opensource.org/licenses/bsd-license.php
"""Contains PackIndexFile and PackFile implementations"""
import zlib
from gitdb.exc import (
BadObject,
AmbiguousObjectName,
UnsupportedOperation,
ParseError
)
from gitdb.util import (
mman,
LazyMixin,
unpack_from,
bin_to_hex,
byte_ord,
)
from gitdb.fun import (
create_pack_object_header,
pack_object_header_info,
is_equal_canonical_sha,
type_id_to_type_map,
write_object,
stream_copy,
chunk_size,
delta_types,
OFS_DELTA,
REF_DELTA,
msb_size
)
try:
from gitdb_speedups._perf import PackIndexFile_sha_to_index
except ImportError:
pass
# END try c module
from gitdb.base import ( # Amazing !
OInfo,
OStream,
OPackInfo,
OPackStream,
ODeltaStream,
ODeltaPackInfo,
ODeltaPackStream,
)
from gitdb.stream import (
DecompressMemMapReader,
DeltaApplyReader,
Sha1Writer,
NullStream,
FlexibleSha1Writer
)
from struct import pack
from binascii import crc32
from gitdb.const import NULL_BYTE
from gitdb.utils.compat import (
izip,
buffer,
xrange,
to_bytes
)
import tempfile
import array
import os
import sys
__all__ = ('PackIndexFile', 'PackFile', 'PackEntity')
#{ Utilities
def pack_object_at(cursor, offset, as_stream):
"""
:return: Tuple(abs_data_offset, PackInfo|PackStream)
an object of the correct type according to the type_id of the object.
If as_stream is True, the object will contain a stream, allowing the
data to be read decompressed.
:param data: random accessible data containing all required information
:parma offset: offset in to the data at which the object information is located
:param as_stream: if True, a stream object will be returned that can read
the data, otherwise you receive an info object only"""
data = cursor.use_region(offset).buffer()
type_id, uncomp_size, data_rela_offset = pack_object_header_info(data)
total_rela_offset = None # set later, actual offset until data stream begins
delta_info = None
# OFFSET DELTA
if type_id == OFS_DELTA:
i = data_rela_offset
c = byte_ord(data[i])
i += 1
delta_offset = c & 0x7f
while c & 0x80:
c = byte_ord(data[i])
i += 1
delta_offset += 1
delta_offset = (delta_offset << 7) + (c & 0x7f)
# END character loop
delta_info = delta_offset
total_rela_offset = i
# REF DELTA
elif type_id == REF_DELTA:
total_rela_offset = data_rela_offset + 20
delta_info = data[data_rela_offset:total_rela_offset]
# BASE OBJECT
else:
# assume its a base object
total_rela_offset = data_rela_offset
# END handle type id
abs_data_offset = offset + total_rela_offset
if as_stream:
stream = DecompressMemMapReader(buffer(data, total_rela_offset), False, uncomp_size)
if delta_info is None:
return abs_data_offset, OPackStream(offset, type_id, uncomp_size, stream)
else:
return abs_data_offset, ODeltaPackStream(offset, type_id, uncomp_size, delta_info, stream)
else:
if delta_info is None:
return abs_data_offset, OPackInfo(offset, type_id, uncomp_size)
else:
return abs_data_offset, ODeltaPackInfo(offset, type_id, uncomp_size, delta_info)
# END handle info
# END handle stream
def write_stream_to_pack(read, write, zstream, base_crc=None):
"""Copy a stream as read from read function, zip it, and write the result.
Count the number of written bytes and return it
:param base_crc: if not None, the crc will be the base for all compressed data
we consecutively write and generate a crc32 from. If None, no crc will be generated
:return: tuple(no bytes read, no bytes written, crc32) crc might be 0 if base_crc
was false"""
br = 0 # bytes read
bw = 0 # bytes written
want_crc = base_crc is not None
crc = 0
if want_crc:
crc = base_crc
# END initialize crc
while True:
chunk = read(chunk_size)
br += len(chunk)
compressed = zstream.compress(chunk)
bw += len(compressed)
write(compressed) # cannot assume return value
if want_crc:
crc = crc32(compressed, crc)
# END handle crc
if len(chunk) != chunk_size:
break
# END copy loop
compressed = zstream.flush()
bw += len(compressed)
write(compressed)
if want_crc:
crc = crc32(compressed, crc)
# END handle crc
return (br, bw, crc)
#} END utilities
class IndexWriter(object):
"""Utility to cache index information, allowing to write all information later
in one go to the given stream
**Note:** currently only writes v2 indices"""
__slots__ = '_objs'
def __init__(self):
self._objs = list()
def append(self, binsha, crc, offset):
"""Append one piece of object information"""
self._objs.append((binsha, crc, offset))
def write(self, pack_sha, write):
"""Write the index file using the given write method
:param pack_sha: binary sha over the whole pack that we index
:return: sha1 binary sha over all index file contents"""
# sort for sha1 hash
self._objs.sort(key=lambda o: o[0])
sha_writer = FlexibleSha1Writer(write)
sha_write = sha_writer.write
sha_write(PackIndexFile.index_v2_signature)
sha_write(pack(">L", PackIndexFile.index_version_default))
# fanout
tmplist = list((0,) * 256) # fanout or list with 64 bit offsets
for t in self._objs:
tmplist[byte_ord(t[0][0])] += 1
# END prepare fanout
for i in xrange(255):
v = tmplist[i]
sha_write(pack('>L', v))
tmplist[i + 1] += v
# END write each fanout entry
sha_write(pack('>L', tmplist[255]))
# sha1 ordered
# save calls, that is push them into c
sha_write(b''.join(t[0] for t in self._objs))
# crc32
for t in self._objs:
sha_write(pack('>L', t[1] & 0xffffffff))
# END for each crc
tmplist = list()
# offset 32
for t in self._objs:
ofs = t[2]
if ofs > 0x7fffffff:
tmplist.append(ofs)
ofs = 0x80000000 + len(tmplist) - 1
# END hande 64 bit offsets
sha_write(pack('>L', ofs & 0xffffffff))
# END for each offset
# offset 64
for ofs in tmplist:
sha_write(pack(">Q", ofs))
# END for each offset
# trailer
assert(len(pack_sha) == 20)
sha_write(pack_sha)
sha = sha_writer.sha(as_hex=False)
write(sha)
return sha
class PackIndexFile(LazyMixin):
"""A pack index provides offsets into the corresponding pack, allowing to find
locations for offsets faster."""
# Dont use slots as we dynamically bind functions for each version, need a dict for this
# The slots you see here are just to keep track of our instance variables
# __slots__ = ('_indexpath', '_fanout_table', '_cursor', '_version',
# '_sha_list_offset', '_crc_list_offset', '_pack_offset', '_pack_64_offset')
# used in v2 indices
_sha_list_offset = 8 + 1024
index_v2_signature = b'\xfftOc'
index_version_default = 2
def __init__(self, indexpath):
super(PackIndexFile, self).__init__()
self._indexpath = indexpath
def close(self):
mman.force_map_handle_removal_win(self._indexpath)
self._cursor = None
def _set_cache_(self, attr):
if attr == "_packfile_checksum":
self._packfile_checksum = self._cursor.map()[-40:-20]
elif attr == "_packfile_checksum":
self._packfile_checksum = self._cursor.map()[-20:]
elif attr == "_cursor":
# Note: We don't lock the file when reading as we cannot be sure
# that we can actually write to the location - it could be a read-only
# alternate for instance
self._cursor = mman.make_cursor(self._indexpath).use_region()
# We will assume that the index will always fully fit into memory !
if mman.window_size() > 0 and self._cursor.file_size() > mman.window_size():
raise AssertionError("The index file at %s is too large to fit into a mapped window (%i > %i). This is a limitation of the implementation" % (
self._indexpath, self._cursor.file_size(), mman.window_size()))
# END assert window size
else:
# now its time to initialize everything - if we are here, someone wants
# to access the fanout table or related properties
# CHECK VERSION
mmap = self._cursor.map()
self._version = (mmap[:4] == self.index_v2_signature and 2) or 1
if self._version == 2:
version_id = unpack_from(">L", mmap, 4)[0]
assert version_id == self._version, "Unsupported index version: %i" % version_id
# END assert version
# SETUP FUNCTIONS
# setup our functions according to the actual version
for fname in ('entry', 'offset', 'sha', 'crc'):
setattr(self, fname, getattr(self, "_%s_v%i" % (fname, self._version)))
# END for each function to initialize
# INITIALIZE DATA
# byte offset is 8 if version is 2, 0 otherwise
self._initialize()
# END handle attributes
#{ Access V1
def _entry_v1(self, i):
""":return: tuple(offset, binsha, 0)"""
return unpack_from(">L20s", self._cursor.map(), 1024 + i * 24) + (0, )
def _offset_v1(self, i):
"""see ``_offset_v2``"""
return unpack_from(">L", self._cursor.map(), 1024 + i * 24)[0]
def _sha_v1(self, i):
"""see ``_sha_v2``"""
base = 1024 + (i * 24) + 4
return self._cursor.map()[base:base + 20]
def _crc_v1(self, i):
"""unsupported"""
return 0
#} END access V1
#{ Access V2
def _entry_v2(self, i):
""":return: tuple(offset, binsha, crc)"""
return (self._offset_v2(i), self._sha_v2(i), self._crc_v2(i))
def _offset_v2(self, i):
""":return: 32 or 64 byte offset into pack files. 64 byte offsets will only
be returned if the pack is larger than 4 GiB, or 2^32"""
offset = unpack_from(">L", self._cursor.map(), self._pack_offset + i * 4)[0]
# if the high-bit is set, this indicates that we have to lookup the offset
# in the 64 bit region of the file. The current offset ( lower 31 bits )
# are the index into it
if offset & 0x80000000:
offset = unpack_from(">Q", self._cursor.map(), self._pack_64_offset + (offset & ~0x80000000) * 8)[0]
# END handle 64 bit offset
return offset
def _sha_v2(self, i):
""":return: sha at the given index of this file index instance"""
base = self._sha_list_offset + i * 20
return self._cursor.map()[base:base + 20]
def _crc_v2(self, i):
""":return: 4 bytes crc for the object at index i"""
return unpack_from(">L", self._cursor.map(), self._crc_list_offset + i * 4)[0]
#} END access V2
#{ Initialization
def _initialize(self):
"""initialize base data"""
self._fanout_table = self._read_fanout((self._version == 2) * 8)
if self._version == 2:
self._crc_list_offset = self._sha_list_offset + self.size() * 20
self._pack_offset = self._crc_list_offset + self.size() * 4
self._pack_64_offset = self._pack_offset + self.size() * 4
# END setup base
def _read_fanout(self, byte_offset):
"""Generate a fanout table from our data"""
d = self._cursor.map()
out = list()
append = out.append
for i in xrange(256):
append(unpack_from('>L', d, byte_offset + i * 4)[0])
# END for each entry
return out
#} END initialization
#{ Properties
def version(self):
return self._version
def size(self):
""":return: amount of objects referred to by this index"""
return self._fanout_table[255]
def path(self):
""":return: path to the packindexfile"""
return self._indexpath
def packfile_checksum(self):
""":return: 20 byte sha representing the sha1 hash of the pack file"""
return self._cursor.map()[-40:-20]
def indexfile_checksum(self):
""":return: 20 byte sha representing the sha1 hash of this index file"""
return self._cursor.map()[-20:]
def offsets(self):
""":return: sequence of all offsets in the order in which they were written
**Note:** return value can be random accessed, but may be immmutable"""
if self._version == 2:
# read stream to array, convert to tuple
a = array.array('I') # 4 byte unsigned int, long are 8 byte on 64 bit it appears
a.fromstring(buffer(self._cursor.map(), self._pack_offset, self._pack_64_offset - self._pack_offset))
# networkbyteorder to something array likes more
if sys.byteorder == 'little':
a.byteswap()
return a
else:
return tuple(self.offset(index) for index in xrange(self.size()))
# END handle version
def sha_to_index(self, sha):
"""
:return: index usable with the ``offset`` or ``entry`` method, or None
if the sha was not found in this pack index
:param sha: 20 byte sha to lookup"""
first_byte = byte_ord(sha[0])
get_sha = self.sha
lo = 0 # lower index, the left bound of the bisection
if first_byte != 0:
lo = self._fanout_table[first_byte - 1]
hi = self._fanout_table[first_byte] # the upper, right bound of the bisection
# bisect until we have the sha
while lo < hi:
mid = (lo + hi) // 2
mid_sha = get_sha(mid)
if sha < mid_sha:
hi = mid
elif sha == mid_sha:
return mid
else:
lo = mid + 1
# END handle midpoint
# END bisect
return None
def partial_sha_to_index(self, partial_bin_sha, canonical_length):
"""
:return: index as in `sha_to_index` or None if the sha was not found in this
index file
:param partial_bin_sha: an at least two bytes of a partial binary sha as bytes
:param canonical_length: length of the original hexadecimal representation of the
given partial binary sha
:raise AmbiguousObjectName:"""
if len(partial_bin_sha) < 2:
raise ValueError("Require at least 2 bytes of partial sha")
assert isinstance(partial_bin_sha, bytes), "partial_bin_sha must be bytes"
first_byte = byte_ord(partial_bin_sha[0])
get_sha = self.sha
lo = 0 # lower index, the left bound of the bisection
if first_byte != 0:
lo = self._fanout_table[first_byte - 1]
hi = self._fanout_table[first_byte] # the upper, right bound of the bisection
# fill the partial to full 20 bytes
filled_sha = partial_bin_sha + NULL_BYTE * (20 - len(partial_bin_sha))
# find lowest
while lo < hi:
mid = (lo + hi) // 2
mid_sha = get_sha(mid)
if filled_sha < mid_sha:
hi = mid
elif filled_sha == mid_sha:
# perfect match
lo = mid
break
else:
lo = mid + 1
# END handle midpoint
# END bisect
if lo < self.size():
cur_sha = get_sha(lo)
if is_equal_canonical_sha(canonical_length, partial_bin_sha, cur_sha):
next_sha = None
if lo + 1 < self.size():
next_sha = get_sha(lo + 1)
if next_sha and next_sha == cur_sha:
raise AmbiguousObjectName(partial_bin_sha)
return lo
# END if we have a match
# END if we found something
return None
if 'PackIndexFile_sha_to_index' in globals():
# NOTE: Its just about 25% faster, the major bottleneck might be the attr
# accesses
def sha_to_index(self, sha):
return PackIndexFile_sha_to_index(self, sha)
# END redefine heavy-hitter with c version
#} END properties
class PackFile(LazyMixin):
"""A pack is a file written according to the Version 2 for git packs
As we currently use memory maps, it could be assumed that the maximum size of
packs therefor is 32 bit on 32 bit systems. On 64 bit systems, this should be
fine though.
**Note:** at some point, this might be implemented using streams as well, or
streams are an alternate path in the case memory maps cannot be created
for some reason - one clearly doesn't want to read 10GB at once in that
case"""
__slots__ = ('_packpath', '_cursor', '_size', '_version')
pack_signature = 0x5041434b # 'PACK'
pack_version_default = 2
# offset into our data at which the first object starts
first_object_offset = 3 * 4 # header bytes
footer_size = 20 # final sha
def __init__(self, packpath):
self._packpath = packpath
def close(self):
mman.force_map_handle_removal_win(self._packpath)
self._cursor = None
def _set_cache_(self, attr):
# we fill the whole cache, whichever attribute gets queried first
self._cursor = mman.make_cursor(self._packpath).use_region()
# read the header information
type_id, self._version, self._size = unpack_from(">LLL", self._cursor.map(), 0)
# TODO: figure out whether we should better keep the lock, or maybe
# add a .keep file instead ?
if type_id != self.pack_signature:
raise ParseError("Invalid pack signature: %i" % type_id)
def _iter_objects(self, start_offset, as_stream=True):
"""Handle the actual iteration of objects within this pack"""
c = self._cursor
content_size = c.file_size() - self.footer_size
cur_offset = start_offset or self.first_object_offset
null = NullStream()
while cur_offset < content_size:
data_offset, ostream = pack_object_at(c, cur_offset, True)
# scrub the stream to the end - this decompresses the object, but yields
# the amount of compressed bytes we need to get to the next offset
stream_copy(ostream.read, null.write, ostream.size, chunk_size)
assert ostream.stream._br == ostream.size
cur_offset += (data_offset - ostream.pack_offset) + ostream.stream.compressed_bytes_read()
# if a stream is requested, reset it beforehand
# Otherwise return the Stream object directly, its derived from the
# info object
if as_stream:
ostream.stream.seek(0)
yield ostream
# END until we have read everything
#{ Pack Information
def size(self):
""":return: The amount of objects stored in this pack"""
return self._size
def version(self):
""":return: the version of this pack"""
return self._version
def data(self):
"""
:return: read-only data of this pack. It provides random access and usually
is a memory map.
:note: This method is unsafe as it returns a window into a file which might be larger than than the actual window size"""
# can use map as we are starting at offset 0. Otherwise we would have to use buffer()
return self._cursor.use_region().map()
def checksum(self):
""":return: 20 byte sha1 hash on all object sha's contained in this file"""
return self._cursor.use_region(self._cursor.file_size() - 20).buffer()[:]
def path(self):
""":return: path to the packfile"""
return self._packpath
#} END pack information
#{ Pack Specific
def collect_streams(self, offset):
"""
:return: list of pack streams which are required to build the object
at the given offset. The first entry of the list is the object at offset,
the last one is either a full object, or a REF_Delta stream. The latter
type needs its reference object to be locked up in an ODB to form a valid
delta chain.
If the object at offset is no delta, the size of the list is 1.
:param offset: specifies the first byte of the object within this pack"""
out = list()
c = self._cursor
while True:
ostream = pack_object_at(c, offset, True)[1]
out.append(ostream)
if ostream.type_id == OFS_DELTA:
offset = ostream.pack_offset - ostream.delta_info
else:
# the only thing we can lookup are OFFSET deltas. Everything
# else is either an object, or a ref delta, in the latter
# case someone else has to find it
break
# END handle type
# END while chaining streams
return out
#} END pack specific
#{ Read-Database like Interface
def info(self, offset):
"""Retrieve information about the object at the given file-absolute offset
:param offset: byte offset
:return: OPackInfo instance, the actual type differs depending on the type_id attribute"""
return pack_object_at(self._cursor, offset or self.first_object_offset, False)[1]
def stream(self, offset):
"""Retrieve an object at the given file-relative offset as stream along with its information
:param offset: byte offset
:return: OPackStream instance, the actual type differs depending on the type_id attribute"""
return pack_object_at(self._cursor, offset or self.first_object_offset, True)[1]
def stream_iter(self, start_offset=0):
"""
:return: iterator yielding OPackStream compatible instances, allowing
to access the data in the pack directly.
:param start_offset: offset to the first object to iterate. If 0, iteration
starts at the very first object in the pack.
**Note:** Iterating a pack directly is costly as the datastream has to be decompressed
to determine the bounds between the objects"""
return self._iter_objects(start_offset, as_stream=True)
#} END Read-Database like Interface
class PackEntity(LazyMixin):
"""Combines the PackIndexFile and the PackFile into one, allowing the
actual objects to be resolved and iterated"""
__slots__ = ('_index', # our index file
'_pack', # our pack file
'_offset_map' # on demand dict mapping one offset to the next consecutive one
)
IndexFileCls = PackIndexFile
PackFileCls = PackFile
def __init__(self, pack_or_index_path):
"""Initialize ourselves with the path to the respective pack or index file"""
basename, ext = os.path.splitext(pack_or_index_path)
self._index = self.IndexFileCls("%s.idx" % basename) # PackIndexFile instance
self._pack = self.PackFileCls("%s.pack" % basename) # corresponding PackFile instance
def close(self):
self._index.close()
self._pack.close()
def _set_cache_(self, attr):
# currently this can only be _offset_map
# TODO: make this a simple sorted offset array which can be bisected
# to find the respective entry, from which we can take a +1 easily
# This might be slower, but should also be much lighter in memory !
offsets_sorted = sorted(self._index.offsets())
last_offset = len(self._pack.data()) - self._pack.footer_size
assert offsets_sorted, "Cannot handle empty indices"
offset_map = None
if len(offsets_sorted) == 1:
offset_map = {offsets_sorted[0]: last_offset}
else:
iter_offsets = iter(offsets_sorted)
iter_offsets_plus_one = iter(offsets_sorted)
next(iter_offsets_plus_one)
consecutive = izip(iter_offsets, iter_offsets_plus_one)
offset_map = dict(consecutive)
# the last offset is not yet set
offset_map[offsets_sorted[-1]] = last_offset
# END handle offset amount
self._offset_map = offset_map
def _sha_to_index(self, sha):
""":return: index for the given sha, or raise"""
index = self._index.sha_to_index(sha)
if index is None:
raise BadObject(sha)
return index
def _iter_objects(self, as_stream):
"""Iterate over all objects in our index and yield their OInfo or OStream instences"""
_sha = self._index.sha
_object = self._object
for index in xrange(self._index.size()):
yield _object(_sha(index), as_stream, index)
# END for each index
def _object(self, sha, as_stream, index=-1):
""":return: OInfo or OStream object providing information about the given sha
:param index: if not -1, its assumed to be the sha's index in the IndexFile"""
# its a little bit redundant here, but it needs to be efficient
if index < 0:
index = self._sha_to_index(sha)
if sha is None:
sha = self._index.sha(index)
# END assure sha is present ( in output )
offset = self._index.offset(index)
type_id, uncomp_size, data_rela_offset = pack_object_header_info(self._pack._cursor.use_region(offset).buffer())
if as_stream:
if type_id not in delta_types:
packstream = self._pack.stream(offset)
return OStream(sha, packstream.type, packstream.size, packstream.stream)
# END handle non-deltas
# produce a delta stream containing all info
# To prevent it from applying the deltas when querying the size,
# we extract it from the delta stream ourselves
streams = self.collect_streams_at_offset(offset)
dstream = DeltaApplyReader.new(streams)
return ODeltaStream(sha, dstream.type, None, dstream)
else:
if type_id not in delta_types:
return OInfo(sha, type_id_to_type_map[type_id], uncomp_size)
# END handle non-deltas
# deltas are a little tougher - unpack the first bytes to obtain
# the actual target size, as opposed to the size of the delta data
streams = self.collect_streams_at_offset(offset)
buf = streams[0].read(512)
offset, src_size = msb_size(buf)
offset, target_size = msb_size(buf, offset)
# collect the streams to obtain the actual object type
if streams[-1].type_id in delta_types:
raise BadObject(sha, "Could not resolve delta object")
return OInfo(sha, streams[-1].type, target_size)
# END handle stream
#{ Read-Database like Interface
def info(self, sha):
"""Retrieve information about the object identified by the given sha
:param sha: 20 byte sha1
:raise BadObject:
:return: OInfo instance, with 20 byte sha"""
return self._object(sha, False)
def stream(self, sha):
"""Retrieve an object stream along with its information as identified by the given sha
:param sha: 20 byte sha1
:raise BadObject:
:return: OStream instance, with 20 byte sha"""
return self._object(sha, True)
def info_at_index(self, index):
"""As ``info``, but uses a PackIndexFile compatible index to refer to the object"""
return self._object(None, False, index)
def stream_at_index(self, index):
"""As ``stream``, but uses a PackIndexFile compatible index to refer to the
object"""
return self._object(None, True, index)
#} END Read-Database like Interface
#{ Interface
def pack(self):
""":return: the underlying pack file instance"""
return self._pack
def index(self):
""":return: the underlying pack index file instance"""
return self._index
def is_valid_stream(self, sha, use_crc=False):
"""
Verify that the stream at the given sha is valid.
:param use_crc: if True, the index' crc is run over the compressed stream of
the object, which is much faster than checking the sha1. It is also
more prone to unnoticed corruption or manipulation.
:param sha: 20 byte sha1 of the object whose stream to verify
whether the compressed stream of the object is valid. If it is
a delta, this only verifies that the delta's data is valid, not the
data of the actual undeltified object, as it depends on more than
just this stream.
If False, the object will be decompressed and the sha generated. It must
match the given sha
:return: True if the stream is valid
:raise UnsupportedOperation: If the index is version 1 only
:raise BadObject: sha was not found"""
if use_crc:
if self._index.version() < 2:
raise UnsupportedOperation("Version 1 indices do not contain crc's, verify by sha instead")
# END handle index version
index = self._sha_to_index(sha)
offset = self._index.offset(index)
next_offset = self._offset_map[offset]
crc_value = self._index.crc(index)
# create the current crc value, on the compressed object data
# Read it in chunks, without copying the data
crc_update = zlib.crc32
pack_data = self._pack.data()
cur_pos = offset
this_crc_value = 0
while cur_pos < next_offset:
rbound = min(cur_pos + chunk_size, next_offset)
size = rbound - cur_pos
this_crc_value = crc_update(buffer(pack_data, cur_pos, size), this_crc_value)
cur_pos += size
# END window size loop
# crc returns signed 32 bit numbers, the AND op forces it into unsigned
# mode ... wow, sneaky, from dulwich.
return (this_crc_value & 0xffffffff) == crc_value
else:
shawriter = Sha1Writer()
stream = self._object(sha, as_stream=True)
# write a loose object, which is the basis for the sha
write_object(stream.type, stream.size, stream.read, shawriter.write)
assert shawriter.sha(as_hex=False) == sha
return shawriter.sha(as_hex=False) == sha
# END handle crc/sha verification
return True
def info_iter(self):
"""
:return: Iterator over all objects in this pack. The iterator yields
OInfo instances"""
return self._iter_objects(as_stream=False)
def stream_iter(self):
"""
:return: iterator over all objects in this pack. The iterator yields
OStream instances"""
return self._iter_objects(as_stream=True)
def collect_streams_at_offset(self, offset):
"""
As the version in the PackFile, but can resolve REF deltas within this pack
For more info, see ``collect_streams``
:param offset: offset into the pack file at which the object can be found"""
streams = self._pack.collect_streams(offset)
# try to resolve the last one if needed. It is assumed to be either
# a REF delta, or a base object, as OFFSET deltas are resolved by the pack
if streams[-1].type_id == REF_DELTA:
stream = streams[-1]
while stream.type_id in delta_types:
if stream.type_id == REF_DELTA:
sindex = self._index.sha_to_index(to_bytes(stream.delta_info))
if sindex is None:
break
stream = self._pack.stream(self._index.offset(sindex))
streams.append(stream)
else:
# must be another OFS DELTA - this could happen if a REF
# delta we resolve previously points to an OFS delta. Who
# would do that ;) ? We can handle it though
stream = self._pack.stream(stream.delta_info)
streams.append(stream)
# END handle ref delta
# END resolve ref streams
# END resolve streams
return streams
def collect_streams(self, sha):
"""
As ``PackFile.collect_streams``, but takes a sha instead of an offset.
Additionally, ref_delta streams will be resolved within this pack.
If this is not possible, the stream will be left alone, hence it is adivsed
to check for unresolved ref-deltas and resolve them before attempting to
construct a delta stream.
:param sha: 20 byte sha1 specifying the object whose related streams you want to collect
:return: list of streams, first being the actual object delta, the last being
a possibly unresolved base object.
:raise BadObject:"""
return self.collect_streams_at_offset(self._index.offset(self._sha_to_index(sha)))
@classmethod
def write_pack(cls, object_iter, pack_write, index_write=None,
object_count=None, zlib_compression=zlib.Z_BEST_SPEED):
"""
Create a new pack by putting all objects obtained by the object_iterator
into a pack which is written using the pack_write method.
The respective index is produced as well if index_write is not Non.
:param object_iter: iterator yielding odb output objects
:param pack_write: function to receive strings to write into the pack stream
:param indx_write: if not None, the function writes the index file corresponding
to the pack.
:param object_count: if you can provide the amount of objects in your iteration,
this would be the place to put it. Otherwise we have to pre-iterate and store
all items into a list to get the number, which uses more memory than necessary.
:param zlib_compression: the zlib compression level to use
:return: tuple(pack_sha, index_binsha) binary sha over all the contents of the pack
and over all contents of the index. If index_write was None, index_binsha will be None
**Note:** The destination of the write functions is up to the user. It could
be a socket, or a file for instance
**Note:** writes only undeltified objects"""
objs = object_iter
if not object_count:
if not isinstance(object_iter, (tuple, list)):
objs = list(object_iter)
# END handle list type
object_count = len(objs)
# END handle object
pack_writer = FlexibleSha1Writer(pack_write)
pwrite = pack_writer.write
ofs = 0 # current offset into the pack file
index = None
wants_index = index_write is not None
# write header
pwrite(pack('>LLL', PackFile.pack_signature, PackFile.pack_version_default, object_count))
ofs += 12
if wants_index:
index = IndexWriter()
# END handle index header
actual_count = 0
for obj in objs:
actual_count += 1
crc = 0
# object header
hdr = create_pack_object_header(obj.type_id, obj.size)
if index_write:
crc = crc32(hdr)
else:
crc = None
# END handle crc
pwrite(hdr)
# data stream
zstream = zlib.compressobj(zlib_compression)
ostream = obj.stream
br, bw, crc = write_stream_to_pack(ostream.read, pwrite, zstream, base_crc=crc)
assert(br == obj.size)
if wants_index:
index.append(obj.binsha, crc, ofs)
# END handle index
ofs += len(hdr) + bw
if actual_count == object_count:
break
# END abort once we are done
# END for each object
if actual_count != object_count:
raise ValueError(
"Expected to write %i objects into pack, but received only %i from iterators" % (object_count, actual_count))
# END count assertion
# write footer
pack_sha = pack_writer.sha(as_hex=False)
assert len(pack_sha) == 20
pack_write(pack_sha)
ofs += len(pack_sha) # just for completeness ;)
index_sha = None
if wants_index:
index_sha = index.write(pack_sha, index_write)
# END handle index
return pack_sha, index_sha
@classmethod
def create(cls, object_iter, base_dir, object_count=None, zlib_compression=zlib.Z_BEST_SPEED):
"""Create a new on-disk entity comprised of a properly named pack file and a properly named
and corresponding index file. The pack contains all OStream objects contained in object iter.
:param base_dir: directory which is to contain the files
:return: PackEntity instance initialized with the new pack
**Note:** for more information on the other parameters see the write_pack method"""
pack_fd, pack_path = tempfile.mkstemp('', 'pack', base_dir)
index_fd, index_path = tempfile.mkstemp('', 'index', base_dir)
pack_write = lambda d: os.write(pack_fd, d)
index_write = lambda d: os.write(index_fd, d)
pack_binsha, index_binsha = cls.write_pack(object_iter, pack_write, index_write, object_count, zlib_compression)
os.close(pack_fd)
os.close(index_fd)
fmt = "pack-%s.%s"
new_pack_path = os.path.join(base_dir, fmt % (bin_to_hex(pack_binsha), 'pack'))
new_index_path = os.path.join(base_dir, fmt % (bin_to_hex(pack_binsha), 'idx'))
os.rename(pack_path, new_pack_path)
os.rename(index_path, new_index_path)
return cls(new_pack_path)
#} END interface
|