mirror of
				https://github.com/superseriousbusiness/gotosocial.git
				synced 2025-10-30 22:42:24 -05:00 
			
		
		
		
	* refactor file handling a tiny bit * whoops * make processing media / emoji defers a bit clear to see that it's the "on finished processing" path * some wording * add some debug logging * add mutex locks for processing remote media * try removing freshness check * fix derefMedia not being allocated * fix log format string * handle case of empty file paths (i.e. not stored) * remove media / emoji once finished processing from dereferencer maps * whoops, fix the cached / force checks * move url parsing outside of 'process___Safely()' funcs to prevalidate url * use emoji.ShortcodeDomain() * update RefreshEmoji() to also match RefreshMedia() changes --------- Co-authored-by: tobi <tobi.smethurst@protonmail.com>
		
			
				
	
	
		
			304 lines
		
	
	
	
		
			9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			304 lines
		
	
	
	
		
			9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // GoToSocial
 | |
| // Copyright (C) GoToSocial Authors admin@gotosocial.org
 | |
| // SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful,
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package storage
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"mime"
 | |
| 	"net/url"
 | |
| 	"os"
 | |
| 	"path"
 | |
| 	"time"
 | |
| 
 | |
| 	"codeberg.org/gruf/go-bytesize"
 | |
| 	"codeberg.org/gruf/go-cache/v3/ttl"
 | |
| 	"codeberg.org/gruf/go-storage"
 | |
| 	"codeberg.org/gruf/go-storage/disk"
 | |
| 	"codeberg.org/gruf/go-storage/s3"
 | |
| 	"github.com/minio/minio-go/v7"
 | |
| 	"github.com/minio/minio-go/v7/pkg/credentials"
 | |
| 	"github.com/superseriousbusiness/gotosocial/internal/config"
 | |
| 	"github.com/superseriousbusiness/gotosocial/internal/gtserror"
 | |
| 	"github.com/superseriousbusiness/gotosocial/internal/log"
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	urlCacheTTL             = time.Hour * 24
 | |
| 	urlCacheExpiryFrequency = time.Minute * 5
 | |
| )
 | |
| 
 | |
| // PresignedURL represents a pre signed S3 URL with
 | |
| // an expiry time.
 | |
| type PresignedURL struct {
 | |
| 	*url.URL
 | |
| 	Expiry time.Time // link expires at this time
 | |
| }
 | |
| 
 | |
| // IsInvalidKey returns whether error is an invalid-key
 | |
| // type error returned by the underlying storage library.
 | |
| func IsInvalidKey(err error) bool {
 | |
| 	return errors.Is(err, storage.ErrInvalidKey)
 | |
| }
 | |
| 
 | |
| // IsAlreadyExist returns whether error is an already-exists
 | |
| // type error returned by the underlying storage library.
 | |
| func IsAlreadyExist(err error) bool {
 | |
| 	return errors.Is(err, storage.ErrAlreadyExists)
 | |
| }
 | |
| 
 | |
| // IsNotFound returns whether error is a not-found error
 | |
| // type returned by the underlying storage library.
 | |
| func IsNotFound(err error) bool {
 | |
| 	return errors.Is(err, storage.ErrNotFound)
 | |
| }
 | |
| 
 | |
| // Driver wraps a kv.KVStore to also provide S3 presigned GET URLs.
 | |
| type Driver struct {
 | |
| 	// Underlying storage
 | |
| 	Storage storage.Storage
 | |
| 
 | |
| 	// S3-only parameters
 | |
| 	Proxy          bool
 | |
| 	Bucket         string
 | |
| 	PresignedCache *ttl.Cache[string, PresignedURL]
 | |
| }
 | |
| 
 | |
| // Get returns the byte value for key in storage.
 | |
| func (d *Driver) Get(ctx context.Context, key string) ([]byte, error) {
 | |
| 	return d.Storage.ReadBytes(ctx, key)
 | |
| }
 | |
| 
 | |
| // GetStream returns an io.ReadCloser for the value bytes at key in the storage.
 | |
| func (d *Driver) GetStream(ctx context.Context, key string) (io.ReadCloser, error) {
 | |
| 	return d.Storage.ReadStream(ctx, key)
 | |
| }
 | |
| 
 | |
| // Put writes the supplied value bytes at key in the storage
 | |
| func (d *Driver) Put(ctx context.Context, key string, value []byte) (int, error) {
 | |
| 	return d.Storage.WriteBytes(ctx, key, value)
 | |
| }
 | |
| 
 | |
| // PutStream writes the bytes from supplied reader at key in the storage
 | |
| func (d *Driver) PutStream(ctx context.Context, key string, r io.Reader) (int64, error) {
 | |
| 	return d.Storage.WriteStream(ctx, key, r)
 | |
| }
 | |
| 
 | |
| // PutFile moves the contents of file at path, to storage.Driver{} under given key.
 | |
| func (d *Driver) PutFile(ctx context.Context, key string, filepath string) (int64, error) {
 | |
| 	// Open file at path for reading.
 | |
| 	file, err := os.Open(filepath)
 | |
| 	if err != nil {
 | |
| 		return 0, gtserror.Newf("error opening file %s: %w", filepath, err)
 | |
| 	}
 | |
| 
 | |
| 	// Write the file data to storage under key. Note
 | |
| 	// that for disk.DiskStorage{} this should end up
 | |
| 	// being a highly optimized Linux sendfile syscall.
 | |
| 	sz, err := d.Storage.WriteStream(ctx, key, file)
 | |
| 	if err != nil {
 | |
| 		err = gtserror.Newf("error writing file %s: %w", key, err)
 | |
| 	}
 | |
| 
 | |
| 	// Close the file: done with it.
 | |
| 	if e := file.Close(); e != nil {
 | |
| 		log.Errorf(ctx, "error closing file %s: %v", filepath, e)
 | |
| 	}
 | |
| 
 | |
| 	return sz, err
 | |
| }
 | |
| 
 | |
| // Delete attempts to remove the supplied key (and corresponding value) from storage.
 | |
| func (d *Driver) Delete(ctx context.Context, key string) error {
 | |
| 	return d.Storage.Remove(ctx, key)
 | |
| }
 | |
| 
 | |
| // Has checks if the supplied key is in the storage.
 | |
| func (d *Driver) Has(ctx context.Context, key string) (bool, error) {
 | |
| 	stat, err := d.Storage.Stat(ctx, key)
 | |
| 	return (stat != nil), err
 | |
| }
 | |
| 
 | |
| // WalkKeys walks the keys in the storage.
 | |
| func (d *Driver) WalkKeys(ctx context.Context, walk func(string) error) error {
 | |
| 	return d.Storage.WalkKeys(ctx, storage.WalkKeysOpts{
 | |
| 		Step: func(entry storage.Entry) error {
 | |
| 			return walk(entry.Key)
 | |
| 		},
 | |
| 	})
 | |
| }
 | |
| 
 | |
| // URL will return a presigned GET object URL, but only if running on S3 storage with proxying disabled.
 | |
| func (d *Driver) URL(ctx context.Context, key string) *PresignedURL {
 | |
| 	// Check whether S3 *without* proxying is enabled
 | |
| 	s3, ok := d.Storage.(*s3.S3Storage)
 | |
| 	if !ok || d.Proxy {
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	// Check cache underlying cache map directly to
 | |
| 	// avoid extending the TTL (which cache.Get() does).
 | |
| 	d.PresignedCache.Lock()
 | |
| 	e, ok := d.PresignedCache.Cache.Get(key)
 | |
| 	d.PresignedCache.Unlock()
 | |
| 
 | |
| 	if ok {
 | |
| 		return &e.Value
 | |
| 	}
 | |
| 
 | |
| 	u, err := s3.Client().PresignedGetObject(ctx, d.Bucket, key, urlCacheTTL, url.Values{
 | |
| 		"response-content-type": []string{mime.TypeByExtension(path.Ext(key))},
 | |
| 	})
 | |
| 	if err != nil {
 | |
| 		// If URL request fails, fallback is to fetch the file. So ignore the error here
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	psu := PresignedURL{
 | |
| 		URL:    u,
 | |
| 		Expiry: time.Now().Add(urlCacheTTL), // link expires in 24h time
 | |
| 	}
 | |
| 
 | |
| 	d.PresignedCache.Set(key, psu)
 | |
| 	return &psu
 | |
| }
 | |
| 
 | |
| // ProbeCSPUri returns a URI string that can be added
 | |
| // to a content-security-policy to allow requests to
 | |
| // endpoints served by this driver.
 | |
| //
 | |
| // If the driver is not backed by non-proxying S3,
 | |
| // this will return an empty string and no error.
 | |
| //
 | |
| // Otherwise, this function probes for a CSP URI by
 | |
| // doing the following:
 | |
| //
 | |
| //  1. Create a temporary file in the S3 bucket.
 | |
| //  2. Generate a pre-signed URL for that file.
 | |
| //  3. Extract '[scheme]://[host]' from the URL.
 | |
| //  4. Remove the temporary file.
 | |
| //  5. Return the '[scheme]://[host]' string.
 | |
| func (d *Driver) ProbeCSPUri(ctx context.Context) (string, error) {
 | |
| 	// Check whether S3 without proxying
 | |
| 	// is enabled. If it's not, there's
 | |
| 	// no need to add anything to the CSP.
 | |
| 	s3, ok := d.Storage.(*s3.S3Storage)
 | |
| 	if !ok || d.Proxy {
 | |
| 		return "", nil
 | |
| 	}
 | |
| 
 | |
| 	const cspKey = "gotosocial-csp-probe"
 | |
| 
 | |
| 	// Create an empty file in S3 storage.
 | |
| 	if _, err := d.Put(ctx, cspKey, make([]byte, 0)); err != nil {
 | |
| 		return "", gtserror.Newf("error putting file in bucket at key %s: %w", cspKey, err)
 | |
| 	}
 | |
| 
 | |
| 	// Try to clean up file whatever happens.
 | |
| 	defer func() {
 | |
| 		if err := d.Delete(ctx, cspKey); err != nil {
 | |
| 			log.Warnf(ctx, "error deleting file from bucket at key %s (%v); "+
 | |
| 				"you may want to remove this file manually from your S3 bucket", cspKey, err)
 | |
| 		}
 | |
| 	}()
 | |
| 
 | |
| 	// Get a presigned URL for that empty file.
 | |
| 	u, err := s3.Client().PresignedGetObject(ctx, d.Bucket, cspKey, 1*time.Second, nil)
 | |
| 	if err != nil {
 | |
| 		return "", err
 | |
| 	}
 | |
| 
 | |
| 	// Create a stripped version of the presigned
 | |
| 	// URL that includes only the host and scheme.
 | |
| 	uStripped := &url.URL{
 | |
| 		Scheme: u.Scheme,
 | |
| 		Host:   u.Host,
 | |
| 	}
 | |
| 
 | |
| 	return uStripped.String(), nil
 | |
| }
 | |
| 
 | |
| func AutoConfig() (*Driver, error) {
 | |
| 	switch backend := config.GetStorageBackend(); backend {
 | |
| 	case "s3":
 | |
| 		return NewS3Storage()
 | |
| 	case "local":
 | |
| 		return NewFileStorage()
 | |
| 	default:
 | |
| 		return nil, fmt.Errorf("invalid storage backend: %s", backend)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func NewFileStorage() (*Driver, error) {
 | |
| 	// Load runtime configuration
 | |
| 	basePath := config.GetStorageLocalBasePath()
 | |
| 
 | |
| 	// Use default disk config with
 | |
| 	// increased write buffer size.
 | |
| 	diskCfg := disk.DefaultConfig()
 | |
| 	diskCfg.WriteBufSize = int(16 * bytesize.KiB)
 | |
| 
 | |
| 	// Open the disk storage implementation
 | |
| 	disk, err := disk.Open(basePath, &diskCfg)
 | |
| 	if err != nil {
 | |
| 		return nil, fmt.Errorf("error opening disk storage: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	return &Driver{
 | |
| 		Storage: disk,
 | |
| 	}, nil
 | |
| }
 | |
| 
 | |
| func NewS3Storage() (*Driver, error) {
 | |
| 	// Load runtime configuration
 | |
| 	endpoint := config.GetStorageS3Endpoint()
 | |
| 	access := config.GetStorageS3AccessKey()
 | |
| 	secret := config.GetStorageS3SecretKey()
 | |
| 	secure := config.GetStorageS3UseSSL()
 | |
| 	bucket := config.GetStorageS3BucketName()
 | |
| 
 | |
| 	// Open the s3 storage implementation
 | |
| 	s3, err := s3.Open(endpoint, bucket, &s3.Config{
 | |
| 		CoreOpts: minio.Options{
 | |
| 			Creds:  credentials.NewStaticV4(access, secret, ""),
 | |
| 			Secure: secure,
 | |
| 		},
 | |
| 		GetOpts:      minio.GetObjectOptions{},
 | |
| 		PutOpts:      minio.PutObjectOptions{},
 | |
| 		PutChunkSize: 5 * 1024 * 1024, // 5MiB
 | |
| 		StatOpts:     minio.StatObjectOptions{},
 | |
| 		RemoveOpts:   minio.RemoveObjectOptions{},
 | |
| 		ListSize:     200,
 | |
| 	})
 | |
| 	if err != nil {
 | |
| 		return nil, fmt.Errorf("error opening s3 storage: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	// ttl should be lower than the expiry used by S3 to avoid serving invalid URLs
 | |
| 	presignedCache := ttl.New[string, PresignedURL](0, 1000, urlCacheTTL-urlCacheExpiryFrequency)
 | |
| 	presignedCache.Start(urlCacheExpiryFrequency)
 | |
| 
 | |
| 	return &Driver{
 | |
| 		Proxy:          config.GetStorageS3Proxy(),
 | |
| 		Bucket:         config.GetStorageS3BucketName(),
 | |
| 		Storage:        s3,
 | |
| 		PresignedCache: presignedCache,
 | |
| 	}, nil
 | |
| }
 |