Change all license headers to comply with REUSE specification. Fix #16132 Co-authored-by: flynnnnnnnnnn <flynnnnnnnnnn@github> Co-authored-by: John Olheiser <john.olheiser@gmail.com>
		
			
				
	
	
		
			211 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			211 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2014 The Gogs Authors. All rights reserved.
 | |
| // SPDX-License-Identifier: MIT
 | |
| 
 | |
| package charset
 | |
| 
 | |
| import (
 | |
| 	"bytes"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"strings"
 | |
| 	"unicode/utf8"
 | |
| 
 | |
| 	"code.gitea.io/gitea/modules/log"
 | |
| 	"code.gitea.io/gitea/modules/setting"
 | |
| 	"code.gitea.io/gitea/modules/util"
 | |
| 
 | |
| 	"github.com/gogs/chardet"
 | |
| 	"golang.org/x/net/html/charset"
 | |
| 	"golang.org/x/text/transform"
 | |
| )
 | |
| 
 | |
| // UTF8BOM is the utf-8 byte-order marker
 | |
| var UTF8BOM = []byte{'\xef', '\xbb', '\xbf'}
 | |
| 
 | |
| // ToUTF8WithFallbackReader detects the encoding of content and converts to UTF-8 reader if possible
 | |
| func ToUTF8WithFallbackReader(rd io.Reader) io.Reader {
 | |
| 	buf := make([]byte, 2048)
 | |
| 	n, err := util.ReadAtMost(rd, buf)
 | |
| 	if err != nil {
 | |
| 		return io.MultiReader(bytes.NewReader(RemoveBOMIfPresent(buf[:n])), rd)
 | |
| 	}
 | |
| 
 | |
| 	charsetLabel, err := DetectEncoding(buf[:n])
 | |
| 	if err != nil || charsetLabel == "UTF-8" {
 | |
| 		return io.MultiReader(bytes.NewReader(RemoveBOMIfPresent(buf[:n])), rd)
 | |
| 	}
 | |
| 
 | |
| 	encoding, _ := charset.Lookup(charsetLabel)
 | |
| 	if encoding == nil {
 | |
| 		return io.MultiReader(bytes.NewReader(buf[:n]), rd)
 | |
| 	}
 | |
| 
 | |
| 	return transform.NewReader(
 | |
| 		io.MultiReader(
 | |
| 			bytes.NewReader(RemoveBOMIfPresent(buf[:n])),
 | |
| 			rd,
 | |
| 		),
 | |
| 		encoding.NewDecoder(),
 | |
| 	)
 | |
| }
 | |
| 
 | |
| // ToUTF8WithErr converts content to UTF8 encoding
 | |
| func ToUTF8WithErr(content []byte) (string, error) {
 | |
| 	charsetLabel, err := DetectEncoding(content)
 | |
| 	if err != nil {
 | |
| 		return "", err
 | |
| 	} else if charsetLabel == "UTF-8" {
 | |
| 		return string(RemoveBOMIfPresent(content)), nil
 | |
| 	}
 | |
| 
 | |
| 	encoding, _ := charset.Lookup(charsetLabel)
 | |
| 	if encoding == nil {
 | |
| 		return string(content), fmt.Errorf("Unknown encoding: %s", charsetLabel)
 | |
| 	}
 | |
| 
 | |
| 	// If there is an error, we concatenate the nicely decoded part and the
 | |
| 	// original left over. This way we won't lose much data.
 | |
| 	result, n, err := transform.Bytes(encoding.NewDecoder(), content)
 | |
| 	if err != nil {
 | |
| 		result = append(result, content[n:]...)
 | |
| 	}
 | |
| 
 | |
| 	result = RemoveBOMIfPresent(result)
 | |
| 
 | |
| 	return string(result), err
 | |
| }
 | |
| 
 | |
| // ToUTF8WithFallback detects the encoding of content and converts to UTF-8 if possible
 | |
| func ToUTF8WithFallback(content []byte) []byte {
 | |
| 	bs, _ := io.ReadAll(ToUTF8WithFallbackReader(bytes.NewReader(content)))
 | |
| 	return bs
 | |
| }
 | |
| 
 | |
| // ToUTF8 converts content to UTF8 encoding and ignore error
 | |
| func ToUTF8(content string) string {
 | |
| 	res, _ := ToUTF8WithErr([]byte(content))
 | |
| 	return res
 | |
| }
 | |
| 
 | |
| // ToUTF8DropErrors makes sure the return string is valid utf-8; attempts conversion if possible
 | |
| func ToUTF8DropErrors(content []byte) []byte {
 | |
| 	charsetLabel, err := DetectEncoding(content)
 | |
| 	if err != nil || charsetLabel == "UTF-8" {
 | |
| 		return RemoveBOMIfPresent(content)
 | |
| 	}
 | |
| 
 | |
| 	encoding, _ := charset.Lookup(charsetLabel)
 | |
| 	if encoding == nil {
 | |
| 		return content
 | |
| 	}
 | |
| 
 | |
| 	// We ignore any non-decodable parts from the file.
 | |
| 	// Some parts might be lost
 | |
| 	var decoded []byte
 | |
| 	decoder := encoding.NewDecoder()
 | |
| 	idx := 0
 | |
| 	for {
 | |
| 		result, n, err := transform.Bytes(decoder, content[idx:])
 | |
| 		decoded = append(decoded, result...)
 | |
| 		if err == nil {
 | |
| 			break
 | |
| 		}
 | |
| 		decoded = append(decoded, ' ')
 | |
| 		idx = idx + n + 1
 | |
| 		if idx >= len(content) {
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return RemoveBOMIfPresent(decoded)
 | |
| }
 | |
| 
 | |
| // RemoveBOMIfPresent removes a UTF-8 BOM from a []byte
 | |
| func RemoveBOMIfPresent(content []byte) []byte {
 | |
| 	if len(content) > 2 && bytes.Equal(content[0:3], UTF8BOM) {
 | |
| 		return content[3:]
 | |
| 	}
 | |
| 	return content
 | |
| }
 | |
| 
 | |
| // DetectEncoding detect the encoding of content
 | |
| func DetectEncoding(content []byte) (string, error) {
 | |
| 	// First we check if the content represents valid utf8 content excepting a truncated character at the end.
 | |
| 
 | |
| 	// Now we could decode all the runes in turn but this is not necessarily the cheapest thing to do
 | |
| 	// instead we walk backwards from the end to trim off a the incomplete character
 | |
| 	toValidate := content
 | |
| 	end := len(toValidate) - 1
 | |
| 
 | |
| 	if end < 0 {
 | |
| 		// no-op
 | |
| 	} else if toValidate[end]>>5 == 0b110 {
 | |
| 		// Incomplete 1 byte extension e.g. © <c2><a9> which has been truncated to <c2>
 | |
| 		toValidate = toValidate[:end]
 | |
| 	} else if end > 0 && toValidate[end]>>6 == 0b10 && toValidate[end-1]>>4 == 0b1110 {
 | |
| 		// Incomplete 2 byte extension e.g. ⛔ <e2><9b><94> which has been truncated to <e2><9b>
 | |
| 		toValidate = toValidate[:end-1]
 | |
| 	} else if end > 1 && toValidate[end]>>6 == 0b10 && toValidate[end-1]>>6 == 0b10 && toValidate[end-2]>>3 == 0b11110 {
 | |
| 		// Incomplete 3 byte extension e.g. 💩 <f0><9f><92><a9> which has been truncated to <f0><9f><92>
 | |
| 		toValidate = toValidate[:end-2]
 | |
| 	}
 | |
| 	if utf8.Valid(toValidate) {
 | |
| 		log.Debug("Detected encoding: utf-8 (fast)")
 | |
| 		return "UTF-8", nil
 | |
| 	}
 | |
| 
 | |
| 	textDetector := chardet.NewTextDetector()
 | |
| 	var detectContent []byte
 | |
| 	if len(content) < 1024 {
 | |
| 		// Check if original content is valid
 | |
| 		if _, err := textDetector.DetectBest(content); err != nil {
 | |
| 			return "", err
 | |
| 		}
 | |
| 		times := 1024 / len(content)
 | |
| 		detectContent = make([]byte, 0, times*len(content))
 | |
| 		for i := 0; i < times; i++ {
 | |
| 			detectContent = append(detectContent, content...)
 | |
| 		}
 | |
| 	} else {
 | |
| 		detectContent = content
 | |
| 	}
 | |
| 
 | |
| 	// Now we can't use DetectBest or just results[0] because the result isn't stable - so we need a tie break
 | |
| 	results, err := textDetector.DetectAll(detectContent)
 | |
| 	if err != nil {
 | |
| 		if err == chardet.NotDetectedError && len(setting.Repository.AnsiCharset) > 0 {
 | |
| 			log.Debug("Using default AnsiCharset: %s", setting.Repository.AnsiCharset)
 | |
| 			return setting.Repository.AnsiCharset, nil
 | |
| 		}
 | |
| 		return "", err
 | |
| 	}
 | |
| 
 | |
| 	topConfidence := results[0].Confidence
 | |
| 	topResult := results[0]
 | |
| 	priority, has := setting.Repository.DetectedCharsetScore[strings.ToLower(strings.TrimSpace(topResult.Charset))]
 | |
| 	for _, result := range results {
 | |
| 		// As results are sorted in confidence order - if we have a different confidence
 | |
| 		// we know it's less than the current confidence and can break out of the loop early
 | |
| 		if result.Confidence != topConfidence {
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 		// Otherwise check if this results is earlier in the DetectedCharsetOrder than our current top guess
 | |
| 		resultPriority, resultHas := setting.Repository.DetectedCharsetScore[strings.ToLower(strings.TrimSpace(result.Charset))]
 | |
| 		if resultHas && (!has || resultPriority < priority) {
 | |
| 			topResult = result
 | |
| 			priority = resultPriority
 | |
| 			has = true
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// FIXME: to properly decouple this function the fallback ANSI charset should be passed as an argument
 | |
| 	if topResult.Charset != "UTF-8" && len(setting.Repository.AnsiCharset) > 0 {
 | |
| 		log.Debug("Using default AnsiCharset: %s", setting.Repository.AnsiCharset)
 | |
| 		return setting.Repository.AnsiCharset, err
 | |
| 	}
 | |
| 
 | |
| 	log.Debug("Detected encoding: %s", topResult.Charset)
 | |
| 	return topResult.Charset, err
 | |
| }
 |