* Dump: Use mholt/archive/v3 to support tar including many compressions Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Dump: Allow dump output to stdout Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Dump: Fixed bug present since #6677 where SessionConfig.Provider is never "file" Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Dump: never pack RepoRootPath, LFS.ContentPath and LogRootPath when they are below AppDataPath Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Dump: also dump LFS (fixes #10058) Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Dump: never dump CustomPath if CustomPath is a subdir of or equal to AppDataPath (fixes #10365) Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * Use log.Info instead of fmt.Fprintf Signed-off-by: Philipp Homann <homann.philipp@googlemail.com> * import ordering * make fmt Co-authored-by: zeripath <art27@cantab.net> Co-authored-by: techknowlogick <techknowlogick@gitea.io> Co-authored-by: Matti R <matti@mdranta.net>
		
			
				
	
	
		
			483 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
		
			Vendored
		
	
	
	
			
		
		
	
	
			483 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
		
			Vendored
		
	
	
	
| // Copyright 2016 The Go Authors. All rights reserved.
 | |
| // Use of this source code is governed by a BSD-style
 | |
| // license that can be found in the LICENSE file.
 | |
| 
 | |
| // +build !appengine
 | |
| // +build gc
 | |
| // +build !noasm
 | |
| 
 | |
| #include "textflag.h"
 | |
| 
 | |
| // The asm code generally follows the pure Go code in decode_other.go, except
 | |
| // where marked with a "!!!".
 | |
| 
 | |
| // func decode(dst, src []byte) int
 | |
| //
 | |
| // All local variables fit into registers. The non-zero stack size is only to
 | |
| // spill registers and push args when issuing a CALL. The register allocation:
 | |
| //	- AX	scratch
 | |
| //	- BX	scratch
 | |
| //	- CX	length or x
 | |
| //	- DX	offset
 | |
| //	- SI	&src[s]
 | |
| //	- DI	&dst[d]
 | |
| //	+ R8	dst_base
 | |
| //	+ R9	dst_len
 | |
| //	+ R10	dst_base + dst_len
 | |
| //	+ R11	src_base
 | |
| //	+ R12	src_len
 | |
| //	+ R13	src_base + src_len
 | |
| //	- R14	used by doCopy
 | |
| //	- R15	used by doCopy
 | |
| //
 | |
| // The registers R8-R13 (marked with a "+") are set at the start of the
 | |
| // function, and after a CALL returns, and are not otherwise modified.
 | |
| //
 | |
| // The d variable is implicitly DI - R8,  and len(dst)-d is R10 - DI.
 | |
| // The s variable is implicitly SI - R11, and len(src)-s is R13 - SI.
 | |
| TEXT ·decode(SB), NOSPLIT, $48-56
 | |
| 	// Initialize SI, DI and R8-R13.
 | |
| 	MOVQ dst_base+0(FP), R8
 | |
| 	MOVQ dst_len+8(FP), R9
 | |
| 	MOVQ R8, DI
 | |
| 	MOVQ R8, R10
 | |
| 	ADDQ R9, R10
 | |
| 	MOVQ src_base+24(FP), R11
 | |
| 	MOVQ src_len+32(FP), R12
 | |
| 	MOVQ R11, SI
 | |
| 	MOVQ R11, R13
 | |
| 	ADDQ R12, R13
 | |
| 
 | |
| loop:
 | |
| 	// for s < len(src)
 | |
| 	CMPQ SI, R13
 | |
| 	JEQ  end
 | |
| 
 | |
| 	// CX = uint32(src[s])
 | |
| 	//
 | |
| 	// switch src[s] & 0x03
 | |
| 	MOVBLZX (SI), CX
 | |
| 	MOVL    CX, BX
 | |
| 	ANDL    $3, BX
 | |
| 	CMPL    BX, $1
 | |
| 	JAE     tagCopy
 | |
| 
 | |
| 	// ----------------------------------------
 | |
| 	// The code below handles literal tags.
 | |
| 
 | |
| 	// case tagLiteral:
 | |
| 	// x := uint32(src[s] >> 2)
 | |
| 	// switch
 | |
| 	SHRL $2, CX
 | |
| 	CMPL CX, $60
 | |
| 	JAE  tagLit60Plus
 | |
| 
 | |
| 	// case x < 60:
 | |
| 	// s++
 | |
| 	INCQ SI
 | |
| 
 | |
| doLit:
 | |
| 	// This is the end of the inner "switch", when we have a literal tag.
 | |
| 	//
 | |
| 	// We assume that CX == x and x fits in a uint32, where x is the variable
 | |
| 	// used in the pure Go decode_other.go code.
 | |
| 
 | |
| 	// length = int(x) + 1
 | |
| 	//
 | |
| 	// Unlike the pure Go code, we don't need to check if length <= 0 because
 | |
| 	// CX can hold 64 bits, so the increment cannot overflow.
 | |
| 	INCQ CX
 | |
| 
 | |
| 	// Prepare to check if copying length bytes will run past the end of dst or
 | |
| 	// src.
 | |
| 	//
 | |
| 	// AX = len(dst) - d
 | |
| 	// BX = len(src) - s
 | |
| 	MOVQ R10, AX
 | |
| 	SUBQ DI, AX
 | |
| 	MOVQ R13, BX
 | |
| 	SUBQ SI, BX
 | |
| 
 | |
| 	// !!! Try a faster technique for short (16 or fewer bytes) copies.
 | |
| 	//
 | |
| 	// if length > 16 || len(dst)-d < 16 || len(src)-s < 16 {
 | |
| 	//   goto callMemmove // Fall back on calling runtime·memmove.
 | |
| 	// }
 | |
| 	//
 | |
| 	// The C++ snappy code calls this TryFastAppend. It also checks len(src)-s
 | |
| 	// against 21 instead of 16, because it cannot assume that all of its input
 | |
| 	// is contiguous in memory and so it needs to leave enough source bytes to
 | |
| 	// read the next tag without refilling buffers, but Go's Decode assumes
 | |
| 	// contiguousness (the src argument is a []byte).
 | |
| 	CMPQ CX, $16
 | |
| 	JGT  callMemmove
 | |
| 	CMPQ AX, $16
 | |
| 	JLT  callMemmove
 | |
| 	CMPQ BX, $16
 | |
| 	JLT  callMemmove
 | |
| 
 | |
| 	// !!! Implement the copy from src to dst as a 16-byte load and store.
 | |
| 	// (Decode's documentation says that dst and src must not overlap.)
 | |
| 	//
 | |
| 	// This always copies 16 bytes, instead of only length bytes, but that's
 | |
| 	// OK. If the input is a valid Snappy encoding then subsequent iterations
 | |
| 	// will fix up the overrun. Otherwise, Decode returns a nil []byte (and a
 | |
| 	// non-nil error), so the overrun will be ignored.
 | |
| 	//
 | |
| 	// Note that on amd64, it is legal and cheap to issue unaligned 8-byte or
 | |
| 	// 16-byte loads and stores. This technique probably wouldn't be as
 | |
| 	// effective on architectures that are fussier about alignment.
 | |
| 	MOVOU 0(SI), X0
 | |
| 	MOVOU X0, 0(DI)
 | |
| 
 | |
| 	// d += length
 | |
| 	// s += length
 | |
| 	ADDQ CX, DI
 | |
| 	ADDQ CX, SI
 | |
| 	JMP  loop
 | |
| 
 | |
| callMemmove:
 | |
| 	// if length > len(dst)-d || length > len(src)-s { etc }
 | |
| 	CMPQ CX, AX
 | |
| 	JGT  errCorrupt
 | |
| 	CMPQ CX, BX
 | |
| 	JGT  errCorrupt
 | |
| 
 | |
| 	// copy(dst[d:], src[s:s+length])
 | |
| 	//
 | |
| 	// This means calling runtime·memmove(&dst[d], &src[s], length), so we push
 | |
| 	// DI, SI and CX as arguments. Coincidentally, we also need to spill those
 | |
| 	// three registers to the stack, to save local variables across the CALL.
 | |
| 	MOVQ DI, 0(SP)
 | |
| 	MOVQ SI, 8(SP)
 | |
| 	MOVQ CX, 16(SP)
 | |
| 	MOVQ DI, 24(SP)
 | |
| 	MOVQ SI, 32(SP)
 | |
| 	MOVQ CX, 40(SP)
 | |
| 	CALL runtime·memmove(SB)
 | |
| 
 | |
| 	// Restore local variables: unspill registers from the stack and
 | |
| 	// re-calculate R8-R13.
 | |
| 	MOVQ 24(SP), DI
 | |
| 	MOVQ 32(SP), SI
 | |
| 	MOVQ 40(SP), CX
 | |
| 	MOVQ dst_base+0(FP), R8
 | |
| 	MOVQ dst_len+8(FP), R9
 | |
| 	MOVQ R8, R10
 | |
| 	ADDQ R9, R10
 | |
| 	MOVQ src_base+24(FP), R11
 | |
| 	MOVQ src_len+32(FP), R12
 | |
| 	MOVQ R11, R13
 | |
| 	ADDQ R12, R13
 | |
| 
 | |
| 	// d += length
 | |
| 	// s += length
 | |
| 	ADDQ CX, DI
 | |
| 	ADDQ CX, SI
 | |
| 	JMP  loop
 | |
| 
 | |
| tagLit60Plus:
 | |
| 	// !!! This fragment does the
 | |
| 	//
 | |
| 	// s += x - 58; if uint(s) > uint(len(src)) { etc }
 | |
| 	//
 | |
| 	// checks. In the asm version, we code it once instead of once per switch case.
 | |
| 	ADDQ CX, SI
 | |
| 	SUBQ $58, SI
 | |
| 	CMPQ SI, R13
 | |
| 	JA   errCorrupt
 | |
| 
 | |
| 	// case x == 60:
 | |
| 	CMPL CX, $61
 | |
| 	JEQ  tagLit61
 | |
| 	JA   tagLit62Plus
 | |
| 
 | |
| 	// x = uint32(src[s-1])
 | |
| 	MOVBLZX -1(SI), CX
 | |
| 	JMP     doLit
 | |
| 
 | |
| tagLit61:
 | |
| 	// case x == 61:
 | |
| 	// x = uint32(src[s-2]) | uint32(src[s-1])<<8
 | |
| 	MOVWLZX -2(SI), CX
 | |
| 	JMP     doLit
 | |
| 
 | |
| tagLit62Plus:
 | |
| 	CMPL CX, $62
 | |
| 	JA   tagLit63
 | |
| 
 | |
| 	// case x == 62:
 | |
| 	// x = uint32(src[s-3]) | uint32(src[s-2])<<8 | uint32(src[s-1])<<16
 | |
| 	MOVWLZX -3(SI), CX
 | |
| 	MOVBLZX -1(SI), BX
 | |
| 	SHLL    $16, BX
 | |
| 	ORL     BX, CX
 | |
| 	JMP     doLit
 | |
| 
 | |
| tagLit63:
 | |
| 	// case x == 63:
 | |
| 	// x = uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24
 | |
| 	MOVL -4(SI), CX
 | |
| 	JMP  doLit
 | |
| 
 | |
| // The code above handles literal tags.
 | |
| // ----------------------------------------
 | |
| // The code below handles copy tags.
 | |
| 
 | |
| tagCopy4:
 | |
| 	// case tagCopy4:
 | |
| 	// s += 5
 | |
| 	ADDQ $5, SI
 | |
| 
 | |
| 	// if uint(s) > uint(len(src)) { etc }
 | |
| 	CMPQ SI, R13
 | |
| 	JA   errCorrupt
 | |
| 
 | |
| 	// length = 1 + int(src[s-5])>>2
 | |
| 	SHRQ $2, CX
 | |
| 	INCQ CX
 | |
| 
 | |
| 	// offset = int(uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24)
 | |
| 	MOVLQZX -4(SI), DX
 | |
| 	JMP     doCopy
 | |
| 
 | |
| tagCopy2:
 | |
| 	// case tagCopy2:
 | |
| 	// s += 3
 | |
| 	ADDQ $3, SI
 | |
| 
 | |
| 	// if uint(s) > uint(len(src)) { etc }
 | |
| 	CMPQ SI, R13
 | |
| 	JA   errCorrupt
 | |
| 
 | |
| 	// length = 1 + int(src[s-3])>>2
 | |
| 	SHRQ $2, CX
 | |
| 	INCQ CX
 | |
| 
 | |
| 	// offset = int(uint32(src[s-2]) | uint32(src[s-1])<<8)
 | |
| 	MOVWQZX -2(SI), DX
 | |
| 	JMP     doCopy
 | |
| 
 | |
| tagCopy:
 | |
| 	// We have a copy tag. We assume that:
 | |
| 	//	- BX == src[s] & 0x03
 | |
| 	//	- CX == src[s]
 | |
| 	CMPQ BX, $2
 | |
| 	JEQ  tagCopy2
 | |
| 	JA   tagCopy4
 | |
| 
 | |
| 	// case tagCopy1:
 | |
| 	// s += 2
 | |
| 	ADDQ $2, SI
 | |
| 
 | |
| 	// if uint(s) > uint(len(src)) { etc }
 | |
| 	CMPQ SI, R13
 | |
| 	JA   errCorrupt
 | |
| 
 | |
| 	// offset = int(uint32(src[s-2])&0xe0<<3 | uint32(src[s-1]))
 | |
| 	MOVQ    CX, DX
 | |
| 	ANDQ    $0xe0, DX
 | |
| 	SHLQ    $3, DX
 | |
| 	MOVBQZX -1(SI), BX
 | |
| 	ORQ     BX, DX
 | |
| 
 | |
| 	// length = 4 + int(src[s-2])>>2&0x7
 | |
| 	SHRQ $2, CX
 | |
| 	ANDQ $7, CX
 | |
| 	ADDQ $4, CX
 | |
| 
 | |
| doCopy:
 | |
| 	// This is the end of the outer "switch", when we have a copy tag.
 | |
| 	//
 | |
| 	// We assume that:
 | |
| 	//	- CX == length && CX > 0
 | |
| 	//	- DX == offset
 | |
| 
 | |
| 	// if offset <= 0 { etc }
 | |
| 	CMPQ DX, $0
 | |
| 	JLE  errCorrupt
 | |
| 
 | |
| 	// if d < offset { etc }
 | |
| 	MOVQ DI, BX
 | |
| 	SUBQ R8, BX
 | |
| 	CMPQ BX, DX
 | |
| 	JLT  errCorrupt
 | |
| 
 | |
| 	// if length > len(dst)-d { etc }
 | |
| 	MOVQ R10, BX
 | |
| 	SUBQ DI, BX
 | |
| 	CMPQ CX, BX
 | |
| 	JGT  errCorrupt
 | |
| 
 | |
| 	// forwardCopy(dst[d:d+length], dst[d-offset:]); d += length
 | |
| 	//
 | |
| 	// Set:
 | |
| 	//	- R14 = len(dst)-d
 | |
| 	//	- R15 = &dst[d-offset]
 | |
| 	MOVQ R10, R14
 | |
| 	SUBQ DI, R14
 | |
| 	MOVQ DI, R15
 | |
| 	SUBQ DX, R15
 | |
| 
 | |
| 	// !!! Try a faster technique for short (16 or fewer bytes) forward copies.
 | |
| 	//
 | |
| 	// First, try using two 8-byte load/stores, similar to the doLit technique
 | |
| 	// above. Even if dst[d:d+length] and dst[d-offset:] can overlap, this is
 | |
| 	// still OK if offset >= 8. Note that this has to be two 8-byte load/stores
 | |
| 	// and not one 16-byte load/store, and the first store has to be before the
 | |
| 	// second load, due to the overlap if offset is in the range [8, 16).
 | |
| 	//
 | |
| 	// if length > 16 || offset < 8 || len(dst)-d < 16 {
 | |
| 	//   goto slowForwardCopy
 | |
| 	// }
 | |
| 	// copy 16 bytes
 | |
| 	// d += length
 | |
| 	CMPQ CX, $16
 | |
| 	JGT  slowForwardCopy
 | |
| 	CMPQ DX, $8
 | |
| 	JLT  slowForwardCopy
 | |
| 	CMPQ R14, $16
 | |
| 	JLT  slowForwardCopy
 | |
| 	MOVQ 0(R15), AX
 | |
| 	MOVQ AX, 0(DI)
 | |
| 	MOVQ 8(R15), BX
 | |
| 	MOVQ BX, 8(DI)
 | |
| 	ADDQ CX, DI
 | |
| 	JMP  loop
 | |
| 
 | |
| slowForwardCopy:
 | |
| 	// !!! If the forward copy is longer than 16 bytes, or if offset < 8, we
 | |
| 	// can still try 8-byte load stores, provided we can overrun up to 10 extra
 | |
| 	// bytes. As above, the overrun will be fixed up by subsequent iterations
 | |
| 	// of the outermost loop.
 | |
| 	//
 | |
| 	// The C++ snappy code calls this technique IncrementalCopyFastPath. Its
 | |
| 	// commentary says:
 | |
| 	//
 | |
| 	// ----
 | |
| 	//
 | |
| 	// The main part of this loop is a simple copy of eight bytes at a time
 | |
| 	// until we've copied (at least) the requested amount of bytes.  However,
 | |
| 	// if d and d-offset are less than eight bytes apart (indicating a
 | |
| 	// repeating pattern of length < 8), we first need to expand the pattern in
 | |
| 	// order to get the correct results. For instance, if the buffer looks like
 | |
| 	// this, with the eight-byte <d-offset> and <d> patterns marked as
 | |
| 	// intervals:
 | |
| 	//
 | |
| 	//    abxxxxxxxxxxxx
 | |
| 	//    [------]           d-offset
 | |
| 	//      [------]         d
 | |
| 	//
 | |
| 	// a single eight-byte copy from <d-offset> to <d> will repeat the pattern
 | |
| 	// once, after which we can move <d> two bytes without moving <d-offset>:
 | |
| 	//
 | |
| 	//    ababxxxxxxxxxx
 | |
| 	//    [------]           d-offset
 | |
| 	//        [------]       d
 | |
| 	//
 | |
| 	// and repeat the exercise until the two no longer overlap.
 | |
| 	//
 | |
| 	// This allows us to do very well in the special case of one single byte
 | |
| 	// repeated many times, without taking a big hit for more general cases.
 | |
| 	//
 | |
| 	// The worst case of extra writing past the end of the match occurs when
 | |
| 	// offset == 1 and length == 1; the last copy will read from byte positions
 | |
| 	// [0..7] and write to [4..11], whereas it was only supposed to write to
 | |
| 	// position 1. Thus, ten excess bytes.
 | |
| 	//
 | |
| 	// ----
 | |
| 	//
 | |
| 	// That "10 byte overrun" worst case is confirmed by Go's
 | |
| 	// TestSlowForwardCopyOverrun, which also tests the fixUpSlowForwardCopy
 | |
| 	// and finishSlowForwardCopy algorithm.
 | |
| 	//
 | |
| 	// if length > len(dst)-d-10 {
 | |
| 	//   goto verySlowForwardCopy
 | |
| 	// }
 | |
| 	SUBQ $10, R14
 | |
| 	CMPQ CX, R14
 | |
| 	JGT  verySlowForwardCopy
 | |
| 
 | |
| makeOffsetAtLeast8:
 | |
| 	// !!! As above, expand the pattern so that offset >= 8 and we can use
 | |
| 	// 8-byte load/stores.
 | |
| 	//
 | |
| 	// for offset < 8 {
 | |
| 	//   copy 8 bytes from dst[d-offset:] to dst[d:]
 | |
| 	//   length -= offset
 | |
| 	//   d      += offset
 | |
| 	//   offset += offset
 | |
| 	//   // The two previous lines together means that d-offset, and therefore
 | |
| 	//   // R15, is unchanged.
 | |
| 	// }
 | |
| 	CMPQ DX, $8
 | |
| 	JGE  fixUpSlowForwardCopy
 | |
| 	MOVQ (R15), BX
 | |
| 	MOVQ BX, (DI)
 | |
| 	SUBQ DX, CX
 | |
| 	ADDQ DX, DI
 | |
| 	ADDQ DX, DX
 | |
| 	JMP  makeOffsetAtLeast8
 | |
| 
 | |
| fixUpSlowForwardCopy:
 | |
| 	// !!! Add length (which might be negative now) to d (implied by DI being
 | |
| 	// &dst[d]) so that d ends up at the right place when we jump back to the
 | |
| 	// top of the loop. Before we do that, though, we save DI to AX so that, if
 | |
| 	// length is positive, copying the remaining length bytes will write to the
 | |
| 	// right place.
 | |
| 	MOVQ DI, AX
 | |
| 	ADDQ CX, DI
 | |
| 
 | |
| finishSlowForwardCopy:
 | |
| 	// !!! Repeat 8-byte load/stores until length <= 0. Ending with a negative
 | |
| 	// length means that we overrun, but as above, that will be fixed up by
 | |
| 	// subsequent iterations of the outermost loop.
 | |
| 	CMPQ CX, $0
 | |
| 	JLE  loop
 | |
| 	MOVQ (R15), BX
 | |
| 	MOVQ BX, (AX)
 | |
| 	ADDQ $8, R15
 | |
| 	ADDQ $8, AX
 | |
| 	SUBQ $8, CX
 | |
| 	JMP  finishSlowForwardCopy
 | |
| 
 | |
| verySlowForwardCopy:
 | |
| 	// verySlowForwardCopy is a simple implementation of forward copy. In C
 | |
| 	// parlance, this is a do/while loop instead of a while loop, since we know
 | |
| 	// that length > 0. In Go syntax:
 | |
| 	//
 | |
| 	// for {
 | |
| 	//   dst[d] = dst[d - offset]
 | |
| 	//   d++
 | |
| 	//   length--
 | |
| 	//   if length == 0 {
 | |
| 	//     break
 | |
| 	//   }
 | |
| 	// }
 | |
| 	MOVB (R15), BX
 | |
| 	MOVB BX, (DI)
 | |
| 	INCQ R15
 | |
| 	INCQ DI
 | |
| 	DECQ CX
 | |
| 	JNZ  verySlowForwardCopy
 | |
| 	JMP  loop
 | |
| 
 | |
| // The code above handles copy tags.
 | |
| // ----------------------------------------
 | |
| 
 | |
| end:
 | |
| 	// This is the end of the "for s < len(src)".
 | |
| 	//
 | |
| 	// if d != len(dst) { etc }
 | |
| 	CMPQ DI, R10
 | |
| 	JNE  errCorrupt
 | |
| 
 | |
| 	// return 0
 | |
| 	MOVQ $0, ret+48(FP)
 | |
| 	RET
 | |
| 
 | |
| errCorrupt:
 | |
| 	// return decodeErrCodeCorrupt
 | |
| 	MOVQ $1, ret+48(FP)
 | |
| 	RET
 |