Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove Python 2 compatibility shims #56

Merged
merged 10 commits into from
Feb 17, 2020
4 changes: 2 additions & 2 deletions gitdb/db/loose.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,11 +50,11 @@
stream_copy
)

from gitdb.utils.compat import MAXSIZE
from gitdb.utils.encoding import force_bytes

import tempfile
import os
import sys


__all__ = ('LooseObjectDB', )
Expand Down Expand Up @@ -196,7 +196,7 @@ def store(self, istream):
if istream.binsha is not None:
# copy as much as possible, the actual uncompressed item size might
# be smaller than the compressed version
stream_copy(istream.read, writer.write, MAXSIZE, self.stream_chunk_size)
stream_copy(istream.read, writer.write, sys.maxsize, self.stream_chunk_size)
else:
# write object with header, we have to make a new one
write_object(istream.type, istream.size, istream.read, writer.write,
Expand Down
3 changes: 1 addition & 2 deletions gitdb/db/pack.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,6 @@
)

from gitdb.pack import PackEntity
from gitdb.utils.compat import xrange

from functools import reduce

Expand Down Expand Up @@ -107,7 +106,7 @@ def sha_iter(self):
for entity in self.entities():
index = entity.index()
sha_by_index = index.sha
for index in xrange(index.size()):
for index in range(index.size()):
yield sha_by_index(index)
# END for each index
# END for each entity
Expand Down
217 changes: 70 additions & 147 deletions gitdb/fun.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,6 @@

from gitdb.const import NULL_BYTE, BYTE_SPACE
from gitdb.utils.encoding import force_text
from gitdb.utils.compat import izip, buffer, xrange, PY3
from gitdb.typ import (
str_blob_type,
str_commit_type,
Expand Down Expand Up @@ -101,7 +100,7 @@ def delta_chunk_apply(dc, bbuf, write):
:param write: write method to call with data to write"""
if dc.data is None:
# COPY DATA FROM SOURCE
write(buffer(bbuf, dc.so, dc.ts))
write(bbuf[dc.so:dc.so + dc.ts])
else:
# APPEND DATA
# whats faster: if + 4 function calls or just a write with a slice ?
Expand Down Expand Up @@ -264,7 +263,7 @@ def compress(self):
# if first_data_index is not None:
nd = StringIO() # new data
so = self[first_data_index].to # start offset in target buffer
for x in xrange(first_data_index, i - 1):
for x in range(first_data_index, i - 1):
xdc = self[x]
nd.write(xdc.data[:xdc.ts])
# END collect data
Expand Down Expand Up @@ -314,7 +313,7 @@ def check_integrity(self, target_size=-1):
right.next()
# this is very pythonic - we might have just use index based access here,
# but this could actually be faster
for lft, rgt in izip(left, right):
for lft, rgt in zip(left, right):
assert lft.rbound() == rgt.to
assert lft.to + lft.ts == rgt.to
# END for each pair
Expand Down Expand Up @@ -424,20 +423,12 @@ def pack_object_header_info(data):
type_id = (c >> 4) & 7 # numeric type
size = c & 15 # starting size
s = 4 # starting bit-shift size
if PY3:
while c & 0x80:
c = byte_ord(data[i])
i += 1
size += (c & 0x7f) << s
s += 7
# END character loop
else:
while c & 0x80:
c = ord(data[i])
i += 1
size += (c & 0x7f) << s
s += 7
# END character loop
while c & 0x80:
c = byte_ord(data[i])
i += 1
size += (c & 0x7f) << s
s += 7
# END character loop
# end performance at expense of maintenance ...
return (type_id, size, i)

Expand All @@ -450,28 +441,16 @@ def create_pack_object_header(obj_type, obj_size):
:param obj_type: pack type_id of the object
:param obj_size: uncompressed size in bytes of the following object stream"""
c = 0 # 1 byte
if PY3:
hdr = bytearray() # output string

c = (obj_type << 4) | (obj_size & 0xf)
obj_size >>= 4
while obj_size:
hdr.append(c | 0x80)
c = obj_size & 0x7f
obj_size >>= 7
# END until size is consumed
hdr.append(c)
else:
hdr = bytes() # output string

c = (obj_type << 4) | (obj_size & 0xf)
obj_size >>= 4
while obj_size:
hdr += chr(c | 0x80)
c = obj_size & 0x7f
obj_size >>= 7
# END until size is consumed
hdr += chr(c)
hdr = bytearray() # output string

c = (obj_type << 4) | (obj_size & 0xf)
obj_size >>= 4
while obj_size:
hdr.append(c | 0x80)
c = obj_size & 0x7f
obj_size >>= 7
# END until size is consumed
hdr.append(c)
# end handle interpreter
return hdr

Expand All @@ -484,26 +463,15 @@ def msb_size(data, offset=0):
i = 0
l = len(data)
hit_msb = False
if PY3:
while i < l:
c = data[i + offset]
size |= (c & 0x7f) << i * 7
i += 1
if not c & 0x80:
hit_msb = True
break
# END check msb bit
# END while in range
else:
while i < l:
c = ord(data[i + offset])
size |= (c & 0x7f) << i * 7
i += 1
if not c & 0x80:
hit_msb = True
break
# END check msb bit
# END while in range
while i < l:
c = data[i + offset]
size |= (c & 0x7f) << i * 7
i += 1
if not c & 0x80:
hit_msb = True
break
# END check msb bit
# END while in range
# end performance ...
if not hit_msb:
raise AssertionError("Could not find terminating MSB byte in data stream")
Expand Down Expand Up @@ -663,93 +631,48 @@ def apply_delta_data(src_buf, src_buf_size, delta_buf, delta_buf_size, write):
**Note:** transcribed to python from the similar routine in patch-delta.c"""
i = 0
db = delta_buf
if PY3:
while i < delta_buf_size:
c = db[i]
i += 1
if c & 0x80:
cp_off, cp_size = 0, 0
if (c & 0x01):
cp_off = db[i]
i += 1
if (c & 0x02):
cp_off |= (db[i] << 8)
i += 1
if (c & 0x04):
cp_off |= (db[i] << 16)
i += 1
if (c & 0x08):
cp_off |= (db[i] << 24)
i += 1
if (c & 0x10):
cp_size = db[i]
i += 1
if (c & 0x20):
cp_size |= (db[i] << 8)
i += 1
if (c & 0x40):
cp_size |= (db[i] << 16)
i += 1

if not cp_size:
cp_size = 0x10000

rbound = cp_off + cp_size
if (rbound < cp_size or
rbound > src_buf_size):
break
write(buffer(src_buf, cp_off, cp_size))
elif c:
write(db[i:i + c])
i += c
else:
raise ValueError("unexpected delta opcode 0")
# END handle command byte
# END while processing delta data
else:
while i < delta_buf_size:
c = ord(db[i])
i += 1
if c & 0x80:
cp_off, cp_size = 0, 0
if (c & 0x01):
cp_off = ord(db[i])
i += 1
if (c & 0x02):
cp_off |= (ord(db[i]) << 8)
i += 1
if (c & 0x04):
cp_off |= (ord(db[i]) << 16)
i += 1
if (c & 0x08):
cp_off |= (ord(db[i]) << 24)
i += 1
if (c & 0x10):
cp_size = ord(db[i])
i += 1
if (c & 0x20):
cp_size |= (ord(db[i]) << 8)
i += 1
if (c & 0x40):
cp_size |= (ord(db[i]) << 16)
i += 1

if not cp_size:
cp_size = 0x10000

rbound = cp_off + cp_size
if (rbound < cp_size or
rbound > src_buf_size):
break
write(buffer(src_buf, cp_off, cp_size))
elif c:
write(db[i:i + c])
i += c
else:
raise ValueError("unexpected delta opcode 0")
# END handle command byte
# END while processing delta data
# end save byte_ord call and prevent performance regression in py2
while i < delta_buf_size:
c = db[i]
i += 1
if c & 0x80:
cp_off, cp_size = 0, 0
if (c & 0x01):
cp_off = db[i]
i += 1
if (c & 0x02):
cp_off |= (db[i] << 8)
i += 1
if (c & 0x04):
cp_off |= (db[i] << 16)
i += 1
if (c & 0x08):
cp_off |= (db[i] << 24)
i += 1
if (c & 0x10):
cp_size = db[i]
i += 1
if (c & 0x20):
cp_size |= (db[i] << 8)
i += 1
if (c & 0x40):
cp_size |= (db[i] << 16)
i += 1

if not cp_size:
cp_size = 0x10000

rbound = cp_off + cp_size
if (rbound < cp_size or
rbound > src_buf_size):
break
write(src_buf[cp_off:cp_off + cp_size])
elif c:
write(db[i:i + c])
i += c
else:
raise ValueError("unexpected delta opcode 0")
# END handle command byte
# END while processing delta data

# yes, lets use the exact same error message that git uses :)
assert i == delta_buf_size, "delta replay has gone wild"
Expand Down
Loading