mirror of https://github.com/aptly-dev/aptly
552 lines
15 KiB
Go
552 lines
15 KiB
Go
package s3
|
||
|
||
import (
|
||
"context"
|
||
"fmt"
|
||
"io"
|
||
"os"
|
||
"path/filepath"
|
||
"strings"
|
||
|
||
"github.com/aptly-dev/aptly/aptly"
|
||
"github.com/aptly-dev/aptly/utils"
|
||
|
||
"github.com/aws/aws-sdk-go-v2/aws"
|
||
v4 "github.com/aws/aws-sdk-go-v2/aws/signer/v4"
|
||
"github.com/aws/aws-sdk-go-v2/config"
|
||
"github.com/aws/aws-sdk-go-v2/credentials"
|
||
"github.com/aws/aws-sdk-go-v2/service/s3"
|
||
"github.com/aws/aws-sdk-go-v2/service/s3/types"
|
||
"github.com/aws/smithy-go"
|
||
"github.com/aws/smithy-go/logging"
|
||
"github.com/pkg/errors"
|
||
"github.com/rs/zerolog"
|
||
"github.com/rs/zerolog/log"
|
||
)
|
||
|
||
const errCodeNotFound = "NotFound"
|
||
|
||
type logger struct{}
|
||
|
||
func (l *logger) Logf(classification logging.Classification, format string, v ...interface{}) {
|
||
var e *zerolog.Event
|
||
switch classification {
|
||
case logging.Debug:
|
||
e = log.Logger.Debug()
|
||
case logging.Warn:
|
||
e = log.Logger.Warn()
|
||
default:
|
||
e = log.Logger.Error()
|
||
}
|
||
e.Msgf(format, v...)
|
||
}
|
||
|
||
// PublishedStorage abstract file system with published files (actually hosted on S3)
|
||
type PublishedStorage struct {
|
||
s3 *s3.Client
|
||
config *aws.Config
|
||
bucket string
|
||
acl types.ObjectCannedACL
|
||
prefix string
|
||
storageClass types.StorageClass
|
||
encryptionMethod types.ServerSideEncryption
|
||
plusWorkaround bool
|
||
disableMultiDel bool
|
||
pathCache map[string]string
|
||
|
||
// True if the bucket encrypts objects by default.
|
||
encryptByDefault bool
|
||
}
|
||
|
||
// Check interface
|
||
var (
|
||
_ aptly.PublishedStorage = (*PublishedStorage)(nil)
|
||
)
|
||
|
||
// NewPublishedStorageRaw creates published storage from raw aws credentials
|
||
func NewPublishedStorageRaw(
|
||
bucket, defaultACL, prefix, storageClass, encryptionMethod string,
|
||
plusWorkaround, disabledMultiDel, forceVirtualHostedStyle bool,
|
||
config *aws.Config, endpoint string,
|
||
) (*PublishedStorage, error) {
|
||
var acl types.ObjectCannedACL
|
||
if defaultACL == "" || defaultACL == "private" {
|
||
acl = types.ObjectCannedACLPrivate
|
||
} else if defaultACL == "public-read" {
|
||
acl = types.ObjectCannedACLPublicRead
|
||
} else if defaultACL == "none" {
|
||
acl = ""
|
||
}
|
||
|
||
if storageClass == string(types.StorageClassStandard) {
|
||
storageClass = ""
|
||
}
|
||
|
||
var baseEndpoint *string
|
||
if endpoint != "" {
|
||
baseEndpoint = aws.String(endpoint)
|
||
}
|
||
|
||
result := &PublishedStorage{
|
||
s3: s3.NewFromConfig(*config, func(o *s3.Options) {
|
||
o.UsePathStyle = !forceVirtualHostedStyle
|
||
o.HTTPSignerV4 = v4.NewSigner()
|
||
o.BaseEndpoint = baseEndpoint
|
||
}),
|
||
bucket: bucket,
|
||
config: config,
|
||
acl: acl,
|
||
prefix: prefix,
|
||
storageClass: types.StorageClass(storageClass),
|
||
encryptionMethod: types.ServerSideEncryption(encryptionMethod),
|
||
plusWorkaround: plusWorkaround,
|
||
disableMultiDel: disabledMultiDel,
|
||
}
|
||
|
||
result.setKMSFlag()
|
||
|
||
return result, nil
|
||
}
|
||
|
||
func (storage *PublishedStorage) setKMSFlag() {
|
||
params := &s3.GetBucketEncryptionInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
}
|
||
output, err := storage.s3.GetBucketEncryption(context.TODO(), params)
|
||
if err != nil {
|
||
return
|
||
}
|
||
|
||
if len(output.ServerSideEncryptionConfiguration.Rules) > 0 &&
|
||
output.ServerSideEncryptionConfiguration.Rules[0].ApplyServerSideEncryptionByDefault.SSEAlgorithm == "aws:kms" {
|
||
storage.encryptByDefault = true
|
||
}
|
||
}
|
||
|
||
// NewPublishedStorage creates new instance of PublishedStorage with specified S3 access
|
||
// keys, region and bucket name
|
||
func NewPublishedStorage(
|
||
accessKey, secretKey, sessionToken, region, endpoint, bucket, defaultACL, prefix, storageClass, encryptionMethod string,
|
||
plusWorkaround, disableMultiDel, _, forceVirtualHostedStyle, debug bool) (*PublishedStorage, error) {
|
||
|
||
opts := []func(*config.LoadOptions) error{config.WithRegion(region)}
|
||
if accessKey != "" {
|
||
opts = append(opts, config.WithCredentialsProvider(credentials.NewStaticCredentialsProvider(accessKey, secretKey, sessionToken)))
|
||
}
|
||
|
||
if debug {
|
||
opts = append(opts, config.WithLogger(&logger{}))
|
||
}
|
||
|
||
config, err := config.LoadDefaultConfig(context.TODO(), opts...)
|
||
if err != nil {
|
||
return nil, err
|
||
}
|
||
|
||
result, err := NewPublishedStorageRaw(bucket, defaultACL, prefix, storageClass,
|
||
encryptionMethod, plusWorkaround, disableMultiDel, forceVirtualHostedStyle, &config, endpoint)
|
||
|
||
return result, err
|
||
}
|
||
|
||
// String
|
||
func (storage *PublishedStorage) String() string {
|
||
return fmt.Sprintf("S3: %s:%s/%s", storage.config.Region, storage.bucket, storage.prefix)
|
||
}
|
||
|
||
// MkDir creates directory recursively under public path
|
||
func (storage *PublishedStorage) MkDir(_ string) error {
|
||
// no op for S3
|
||
return nil
|
||
}
|
||
|
||
// PutFile puts file into published storage at specified path
|
||
func (storage *PublishedStorage) PutFile(path string, sourceFilename string) error {
|
||
var (
|
||
source *os.File
|
||
err error
|
||
)
|
||
source, err = os.Open(sourceFilename)
|
||
if err != nil {
|
||
return err
|
||
}
|
||
defer source.Close()
|
||
|
||
err = storage.putFile(path, source, "")
|
||
if err != nil {
|
||
err = errors.Wrap(err, fmt.Sprintf("error uploading %s to %s", sourceFilename, storage))
|
||
}
|
||
|
||
return err
|
||
}
|
||
|
||
// getMD5 retrieves MD5 stored in the metadata, if any
|
||
func (storage *PublishedStorage) getMD5(path string) (string, error) {
|
||
params := &s3.HeadObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path)),
|
||
}
|
||
output, err := storage.s3.HeadObject(context.TODO(), params)
|
||
if err != nil {
|
||
return "", err
|
||
}
|
||
|
||
return output.Metadata["Md5"], nil
|
||
}
|
||
|
||
// putFile uploads file-like object to
|
||
func (storage *PublishedStorage) putFile(path string, source io.ReadSeeker, sourceMD5 string) error {
|
||
|
||
params := &s3.PutObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path)),
|
||
Body: source,
|
||
ACL: storage.acl,
|
||
}
|
||
if storage.storageClass != "" {
|
||
params.StorageClass = types.StorageClass(storage.storageClass)
|
||
}
|
||
if storage.encryptionMethod != "" {
|
||
params.ServerSideEncryption = types.ServerSideEncryption(storage.encryptionMethod)
|
||
}
|
||
if sourceMD5 != "" {
|
||
params.Metadata = map[string]string{
|
||
"Md5": sourceMD5,
|
||
}
|
||
}
|
||
|
||
_, err := storage.s3.PutObject(context.TODO(), params)
|
||
if err != nil {
|
||
return err
|
||
}
|
||
|
||
if storage.plusWorkaround && strings.Contains(path, "+") {
|
||
_, err = source.Seek(0, 0)
|
||
if err != nil {
|
||
return err
|
||
}
|
||
|
||
return storage.putFile(strings.Replace(path, "+", " ", -1), source, sourceMD5)
|
||
}
|
||
return nil
|
||
}
|
||
|
||
// Remove removes single file under public path
|
||
func (storage *PublishedStorage) Remove(path string) error {
|
||
params := &s3.DeleteObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path)),
|
||
}
|
||
|
||
if _, err := storage.s3.DeleteObject(context.TODO(), params); err != nil {
|
||
var notFoundErr *smithy.GenericAPIError
|
||
if errors.As(err, ¬FoundErr) && notFoundErr.Code == "NoSuchBucket" {
|
||
// ignore 'no such bucket' errors on removal
|
||
return nil
|
||
}
|
||
return errors.Wrap(err, fmt.Sprintf("error deleting %s from %s", path, storage))
|
||
}
|
||
|
||
if storage.plusWorkaround && strings.Contains(path, "+") {
|
||
// try to remove workaround version, but don't care about result
|
||
_ = storage.Remove(strings.Replace(path, "+", " ", -1))
|
||
}
|
||
|
||
delete(storage.pathCache, path)
|
||
|
||
return nil
|
||
}
|
||
|
||
// RemoveDirs removes directory structure under public path
|
||
func (storage *PublishedStorage) RemoveDirs(path string, _ aptly.Progress) error {
|
||
const page = 1000
|
||
|
||
filelist, _, err := storage.internalFilelist(path, false)
|
||
if err != nil {
|
||
if errors.Is(err, &types.NoSuchBucket{}) {
|
||
// ignore 'no such bucket' errors on removal
|
||
return nil
|
||
}
|
||
return err
|
||
}
|
||
|
||
if storage.disableMultiDel {
|
||
for i := range filelist {
|
||
params := &s3.DeleteObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path, filelist[i])),
|
||
}
|
||
_, err := storage.s3.DeleteObject(context.TODO(), params)
|
||
if err != nil {
|
||
return fmt.Errorf("error deleting path %s from %s: %s", filelist[i], storage, err)
|
||
}
|
||
delete(storage.pathCache, filepath.Join(path, filelist[i]))
|
||
}
|
||
} else {
|
||
numParts := (len(filelist) + page - 1) / page
|
||
|
||
for i := 0; i < numParts; i++ {
|
||
var part []string
|
||
if i == numParts-1 {
|
||
part = filelist[i*page:]
|
||
} else {
|
||
part = filelist[i*page : (i+1)*page]
|
||
}
|
||
|
||
paths := make([]types.ObjectIdentifier, len(part))
|
||
for i := range part {
|
||
paths[i] = types.ObjectIdentifier{
|
||
Key: aws.String(filepath.Join(storage.prefix, path, part[i])),
|
||
}
|
||
}
|
||
|
||
params := &s3.DeleteObjectsInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Delete: &types.Delete{
|
||
Objects: paths,
|
||
Quiet: true,
|
||
},
|
||
}
|
||
|
||
_, err := storage.s3.DeleteObjects(context.TODO(), params)
|
||
if err != nil {
|
||
return fmt.Errorf("error deleting multiple paths from %s: %s", storage, err)
|
||
}
|
||
for i := range part {
|
||
delete(storage.pathCache, filepath.Join(path, part[i]))
|
||
}
|
||
}
|
||
}
|
||
|
||
return nil
|
||
}
|
||
|
||
// LinkFromPool links package file from pool to dist's pool location
|
||
//
|
||
// publishedPrefix is desired prefix for the location in the pool.
|
||
// publishedRelPath is desired location in pool (like pool/component/liba/libav/)
|
||
// sourcePool is instance of aptly.PackagePool
|
||
// sourcePath is filepath to package file in package pool
|
||
//
|
||
// LinkFromPool returns relative path for the published file to be included in package index
|
||
func (storage *PublishedStorage) LinkFromPool(publishedPrefix, publishedRelPath, fileName string, sourcePool aptly.PackagePool,
|
||
sourcePath string, sourceChecksums utils.ChecksumInfo, force bool) error {
|
||
|
||
publishedDirectory := filepath.Join(publishedPrefix, publishedRelPath)
|
||
relPath := filepath.Join(publishedDirectory, fileName)
|
||
poolPath := filepath.Join(storage.prefix, relPath)
|
||
|
||
if storage.pathCache == nil {
|
||
paths, md5s, err := storage.internalFilelist(filepath.Join(storage.prefix, publishedPrefix, "pool"), true)
|
||
if err != nil {
|
||
return errors.Wrap(err, "error caching paths under prefix")
|
||
}
|
||
|
||
storage.pathCache = make(map[string]string, len(paths))
|
||
|
||
for i := range paths {
|
||
storage.pathCache[paths[i]] = md5s[i]
|
||
}
|
||
}
|
||
|
||
destinationMD5, exists := storage.pathCache[relPath]
|
||
sourceMD5 := sourceChecksums.MD5
|
||
|
||
if exists {
|
||
if sourceMD5 == "" {
|
||
return fmt.Errorf("unable to compare object, MD5 checksum missing")
|
||
}
|
||
|
||
if len(destinationMD5) != 32 || storage.encryptByDefault {
|
||
// doesn’t look like a valid MD5,
|
||
// attempt to fetch one from the metadata
|
||
var err error
|
||
destinationMD5, err = storage.getMD5(relPath)
|
||
if err != nil {
|
||
err = errors.Wrap(err, fmt.Sprintf("error verifying MD5 for %s: %s", storage, poolPath))
|
||
return err
|
||
}
|
||
storage.pathCache[relPath] = destinationMD5
|
||
}
|
||
|
||
if destinationMD5 == sourceMD5 {
|
||
return nil
|
||
}
|
||
|
||
if !force {
|
||
return fmt.Errorf("error putting file to %s: file already exists and is different: %s", poolPath, storage)
|
||
}
|
||
}
|
||
|
||
source, err := sourcePool.Open(sourcePath)
|
||
if err != nil {
|
||
return err
|
||
}
|
||
defer source.Close()
|
||
|
||
err = storage.putFile(relPath, source, sourceMD5)
|
||
if err == nil {
|
||
storage.pathCache[relPath] = sourceMD5
|
||
} else {
|
||
err = errors.Wrap(err, fmt.Sprintf("error uploading %s to %s: %s", sourcePath, storage, poolPath))
|
||
}
|
||
|
||
return err
|
||
}
|
||
|
||
// Filelist returns list of files under prefix
|
||
func (storage *PublishedStorage) Filelist(prefix string) ([]string, error) {
|
||
paths, _, err := storage.internalFilelist(prefix, true)
|
||
return paths, err
|
||
}
|
||
|
||
func (storage *PublishedStorage) internalFilelist(prefix string, hidePlusWorkaround bool) (paths []string, md5s []string, err error) {
|
||
paths = make([]string, 0, 1024)
|
||
md5s = make([]string, 0, 1024)
|
||
prefix = filepath.Join(storage.prefix, prefix)
|
||
if prefix != "" {
|
||
prefix += "/"
|
||
}
|
||
|
||
params := &s3.ListObjectsV2Input{
|
||
Bucket: aws.String(storage.bucket),
|
||
Prefix: aws.String(prefix),
|
||
MaxKeys: 1000,
|
||
}
|
||
|
||
p := s3.NewListObjectsV2Paginator(storage.s3, params)
|
||
for i := 1; p.HasMorePages(); i++ {
|
||
page, err := p.NextPage(context.TODO())
|
||
if err != nil {
|
||
return nil, nil, fmt.Errorf("failed to get page %d: %w", i, err)
|
||
}
|
||
|
||
// Log the objects found
|
||
for _, key := range page.Contents {
|
||
if storage.plusWorkaround && hidePlusWorkaround && strings.Contains(*key.Key, " ") {
|
||
// if we use plusWorkaround, we want to hide those duplicates
|
||
/// from listing
|
||
continue
|
||
}
|
||
|
||
if prefix == "" {
|
||
paths = append(paths, *key.Key)
|
||
} else {
|
||
paths = append(paths, (*key.Key)[len(prefix):])
|
||
}
|
||
md5s = append(md5s, strings.Replace(*key.ETag, "\"", "", -1))
|
||
}
|
||
}
|
||
|
||
if err != nil {
|
||
return nil, nil, errors.WithMessagef(err, "error listing under prefix %s in %s: %s", prefix, storage, err)
|
||
}
|
||
|
||
return paths, md5s, nil
|
||
}
|
||
|
||
// RenameFile renames (moves) file
|
||
func (storage *PublishedStorage) RenameFile(oldName, newName string) error {
|
||
source := fmt.Sprintf("/%s/%s", storage.bucket, filepath.Join(storage.prefix, oldName))
|
||
|
||
params := &s3.CopyObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
CopySource: aws.String(source),
|
||
Key: aws.String(filepath.Join(storage.prefix, newName)),
|
||
ACL: storage.acl,
|
||
}
|
||
|
||
if storage.storageClass != "" {
|
||
params.StorageClass = storage.storageClass
|
||
}
|
||
if storage.encryptionMethod != "" {
|
||
params.ServerSideEncryption = storage.encryptionMethod
|
||
}
|
||
|
||
_, err := storage.s3.CopyObject(context.TODO(), params)
|
||
if err != nil {
|
||
return fmt.Errorf("error copying %s -> %s in %s: %s", oldName, newName, storage, err)
|
||
}
|
||
|
||
return storage.Remove(oldName)
|
||
}
|
||
|
||
// SymLink creates a copy of src file and adds link information as meta data
|
||
func (storage *PublishedStorage) SymLink(src string, dst string) error {
|
||
|
||
params := &s3.CopyObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
CopySource: aws.String(filepath.Join(storage.bucket, storage.prefix, src)),
|
||
Key: aws.String(filepath.Join(storage.prefix, dst)),
|
||
ACL: types.ObjectCannedACL(storage.acl),
|
||
Metadata: map[string]string{
|
||
"SymLink": src,
|
||
},
|
||
MetadataDirective: types.MetadataDirective("REPLACE"),
|
||
}
|
||
|
||
if storage.storageClass != "" {
|
||
params.StorageClass = types.StorageClass(storage.storageClass)
|
||
}
|
||
if storage.encryptionMethod != "" {
|
||
params.ServerSideEncryption = types.ServerSideEncryption(storage.encryptionMethod)
|
||
}
|
||
|
||
_, err := storage.s3.CopyObject(context.TODO(), params)
|
||
if err != nil {
|
||
return fmt.Errorf("error symlinking %s -> %s in %s: %s", src, dst, storage, err)
|
||
}
|
||
|
||
return err
|
||
}
|
||
|
||
// HardLink using symlink functionality as hard links do not exist
|
||
func (storage *PublishedStorage) HardLink(src string, dst string) error {
|
||
return storage.SymLink(src, dst)
|
||
}
|
||
|
||
// FileExists returns true if path exists
|
||
func (storage *PublishedStorage) FileExists(path string) (bool, error) {
|
||
params := &s3.HeadObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path)),
|
||
}
|
||
_, err := storage.s3.HeadObject(context.TODO(), params)
|
||
if err != nil {
|
||
var notFoundErr *types.NotFound
|
||
if errors.As(err, ¬FoundErr) {
|
||
return false, nil
|
||
}
|
||
|
||
// falback in case the above condidition fails
|
||
var opErr *smithy.OperationError
|
||
if errors.As(err, &opErr) {
|
||
var ae smithy.APIError
|
||
if errors.As(err, &ae) {
|
||
if ae.ErrorCode() == "NotFound" {
|
||
return false, nil
|
||
}
|
||
}
|
||
}
|
||
|
||
return false, err
|
||
}
|
||
|
||
return true, nil
|
||
}
|
||
|
||
// ReadLink returns the symbolic link pointed to by path.
|
||
// This simply reads text file created with SymLink
|
||
func (storage *PublishedStorage) ReadLink(path string) (string, error) {
|
||
params := &s3.HeadObjectInput{
|
||
Bucket: aws.String(storage.bucket),
|
||
Key: aws.String(filepath.Join(storage.prefix, path)),
|
||
}
|
||
output, err := storage.s3.HeadObject(context.TODO(), params)
|
||
if err != nil {
|
||
return "", err
|
||
}
|
||
|
||
return output.Metadata["SymLink"], nil
|
||
}
|