mirror of
https://github.com/rocky-linux/peridot.git
synced 2024-11-27 15:36:25 +00:00
1644 lines
54 KiB
Go
1644 lines
54 KiB
Go
|
// Copyright 2014 Google LLC
|
||
|
//
|
||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
// you may not use this file except in compliance with the License.
|
||
|
// You may obtain a copy of the License at
|
||
|
//
|
||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||
|
//
|
||
|
// Unless required by applicable law or agreed to in writing, software
|
||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
// See the License for the specific language governing permissions and
|
||
|
// limitations under the License.
|
||
|
|
||
|
package storage
|
||
|
|
||
|
import (
|
||
|
"bytes"
|
||
|
"context"
|
||
|
"crypto"
|
||
|
"crypto/rand"
|
||
|
"crypto/rsa"
|
||
|
"crypto/sha256"
|
||
|
"crypto/x509"
|
||
|
"encoding/base64"
|
||
|
"encoding/hex"
|
||
|
"encoding/pem"
|
||
|
"errors"
|
||
|
"fmt"
|
||
|
"net/http"
|
||
|
"net/url"
|
||
|
"os"
|
||
|
"reflect"
|
||
|
"regexp"
|
||
|
"sort"
|
||
|
"strconv"
|
||
|
"strings"
|
||
|
"time"
|
||
|
"unicode/utf8"
|
||
|
|
||
|
"cloud.google.com/go/internal/optional"
|
||
|
"cloud.google.com/go/internal/trace"
|
||
|
"cloud.google.com/go/internal/version"
|
||
|
"google.golang.org/api/googleapi"
|
||
|
"google.golang.org/api/option"
|
||
|
raw "google.golang.org/api/storage/v1"
|
||
|
htransport "google.golang.org/api/transport/http"
|
||
|
)
|
||
|
|
||
|
// Methods which can be used in signed URLs.
|
||
|
var signedURLMethods = map[string]bool{"DELETE": true, "GET": true, "HEAD": true, "POST": true, "PUT": true}
|
||
|
|
||
|
var (
|
||
|
// ErrBucketNotExist indicates that the bucket does not exist.
|
||
|
ErrBucketNotExist = errors.New("storage: bucket doesn't exist")
|
||
|
// ErrObjectNotExist indicates that the object does not exist.
|
||
|
ErrObjectNotExist = errors.New("storage: object doesn't exist")
|
||
|
// errMethodNotValid indicates that given HTTP method is not valid.
|
||
|
errMethodNotValid = fmt.Errorf("storage: HTTP method should be one of %v", reflect.ValueOf(signedURLMethods).MapKeys())
|
||
|
)
|
||
|
|
||
|
var userAgent = fmt.Sprintf("gcloud-golang-storage/%s", version.Repo)
|
||
|
|
||
|
const (
|
||
|
// ScopeFullControl grants permissions to manage your
|
||
|
// data and permissions in Google Cloud Storage.
|
||
|
ScopeFullControl = raw.DevstorageFullControlScope
|
||
|
|
||
|
// ScopeReadOnly grants permissions to
|
||
|
// view your data in Google Cloud Storage.
|
||
|
ScopeReadOnly = raw.DevstorageReadOnlyScope
|
||
|
|
||
|
// ScopeReadWrite grants permissions to manage your
|
||
|
// data in Google Cloud Storage.
|
||
|
ScopeReadWrite = raw.DevstorageReadWriteScope
|
||
|
)
|
||
|
|
||
|
var xGoogHeader = fmt.Sprintf("gl-go/%s gccl/%s", version.Go(), version.Repo)
|
||
|
|
||
|
func setClientHeader(headers http.Header) {
|
||
|
headers.Set("x-goog-api-client", xGoogHeader)
|
||
|
}
|
||
|
|
||
|
// Client is a client for interacting with Google Cloud Storage.
|
||
|
//
|
||
|
// Clients should be reused instead of created as needed.
|
||
|
// The methods of Client are safe for concurrent use by multiple goroutines.
|
||
|
type Client struct {
|
||
|
hc *http.Client
|
||
|
raw *raw.Service
|
||
|
// Scheme describes the scheme under the current host.
|
||
|
scheme string
|
||
|
// EnvHost is the host set on the STORAGE_EMULATOR_HOST variable.
|
||
|
envHost string
|
||
|
// ReadHost is the default host used on the reader.
|
||
|
readHost string
|
||
|
}
|
||
|
|
||
|
// NewClient creates a new Google Cloud Storage client.
|
||
|
// The default scope is ScopeFullControl. To use a different scope, like
|
||
|
// ScopeReadOnly, use option.WithScopes.
|
||
|
//
|
||
|
// Clients should be reused instead of created as needed. The methods of Client
|
||
|
// are safe for concurrent use by multiple goroutines.
|
||
|
func NewClient(ctx context.Context, opts ...option.ClientOption) (*Client, error) {
|
||
|
var host, readHost, scheme string
|
||
|
|
||
|
if host = os.Getenv("STORAGE_EMULATOR_HOST"); host == "" {
|
||
|
scheme = "https"
|
||
|
readHost = "storage.googleapis.com"
|
||
|
|
||
|
// Prepend default options to avoid overriding options passed by the user.
|
||
|
opts = append([]option.ClientOption{option.WithScopes(ScopeFullControl), option.WithUserAgent(userAgent)}, opts...)
|
||
|
} else {
|
||
|
scheme = "http"
|
||
|
readHost = host
|
||
|
|
||
|
opts = append([]option.ClientOption{option.WithoutAuthentication()}, opts...)
|
||
|
}
|
||
|
|
||
|
hc, ep, err := htransport.NewClient(ctx, opts...)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf("dialing: %v", err)
|
||
|
}
|
||
|
rawService, err := raw.NewService(ctx, option.WithHTTPClient(hc))
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf("storage client: %v", err)
|
||
|
}
|
||
|
if ep == "" {
|
||
|
// Override the default value for BasePath from the raw client.
|
||
|
// TODO: remove when the raw client uses this endpoint as its default (~end of 2020)
|
||
|
rawService.BasePath = "https://storage.googleapis.com/storage/v1/"
|
||
|
} else {
|
||
|
// If the endpoint has been set explicitly, use this for the BasePath
|
||
|
// as well as readHost
|
||
|
rawService.BasePath = ep
|
||
|
u, err := url.Parse(ep)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf("supplied endpoint %v is not valid: %v", ep, err)
|
||
|
}
|
||
|
readHost = u.Host
|
||
|
}
|
||
|
|
||
|
return &Client{
|
||
|
hc: hc,
|
||
|
raw: rawService,
|
||
|
scheme: scheme,
|
||
|
envHost: host,
|
||
|
readHost: readHost,
|
||
|
}, nil
|
||
|
}
|
||
|
|
||
|
// Close closes the Client.
|
||
|
//
|
||
|
// Close need not be called at program exit.
|
||
|
func (c *Client) Close() error {
|
||
|
// Set fields to nil so that subsequent uses will panic.
|
||
|
c.hc = nil
|
||
|
c.raw = nil
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// SigningScheme determines the API version to use when signing URLs.
|
||
|
type SigningScheme int
|
||
|
|
||
|
const (
|
||
|
// SigningSchemeDefault is presently V2 and will change to V4 in the future.
|
||
|
SigningSchemeDefault SigningScheme = iota
|
||
|
|
||
|
// SigningSchemeV2 uses the V2 scheme to sign URLs.
|
||
|
SigningSchemeV2
|
||
|
|
||
|
// SigningSchemeV4 uses the V4 scheme to sign URLs.
|
||
|
SigningSchemeV4
|
||
|
)
|
||
|
|
||
|
// URLStyle determines the style to use for the signed URL. pathStyle is the
|
||
|
// default. All non-default options work with V4 scheme only. See
|
||
|
// https://cloud.google.com/storage/docs/request-endpoints for details.
|
||
|
type URLStyle interface {
|
||
|
// host should return the host portion of the signed URL, not including
|
||
|
// the scheme (e.g. storage.googleapis.com).
|
||
|
host(bucket string) string
|
||
|
|
||
|
// path should return the path portion of the signed URL, which may include
|
||
|
// both the bucket and object name or only the object name depending on the
|
||
|
// style.
|
||
|
path(bucket, object string) string
|
||
|
}
|
||
|
|
||
|
type pathStyle struct{}
|
||
|
|
||
|
type virtualHostedStyle struct{}
|
||
|
|
||
|
type bucketBoundHostname struct {
|
||
|
hostname string
|
||
|
}
|
||
|
|
||
|
func (s pathStyle) host(bucket string) string {
|
||
|
return "storage.googleapis.com"
|
||
|
}
|
||
|
|
||
|
func (s virtualHostedStyle) host(bucket string) string {
|
||
|
return bucket + ".storage.googleapis.com"
|
||
|
}
|
||
|
|
||
|
func (s bucketBoundHostname) host(bucket string) string {
|
||
|
return s.hostname
|
||
|
}
|
||
|
|
||
|
func (s pathStyle) path(bucket, object string) string {
|
||
|
p := bucket
|
||
|
if object != "" {
|
||
|
p += "/" + object
|
||
|
}
|
||
|
return p
|
||
|
}
|
||
|
|
||
|
func (s virtualHostedStyle) path(bucket, object string) string {
|
||
|
return object
|
||
|
}
|
||
|
|
||
|
func (s bucketBoundHostname) path(bucket, object string) string {
|
||
|
return object
|
||
|
}
|
||
|
|
||
|
// PathStyle is the default style, and will generate a URL of the form
|
||
|
// "storage.googleapis.com/<bucket-name>/<object-name>".
|
||
|
func PathStyle() URLStyle {
|
||
|
return pathStyle{}
|
||
|
}
|
||
|
|
||
|
// VirtualHostedStyle generates a URL relative to the bucket's virtual
|
||
|
// hostname, e.g. "<bucket-name>.storage.googleapis.com/<object-name>".
|
||
|
func VirtualHostedStyle() URLStyle {
|
||
|
return virtualHostedStyle{}
|
||
|
}
|
||
|
|
||
|
// BucketBoundHostname generates a URL with a custom hostname tied to a
|
||
|
// specific GCS bucket. The desired hostname should be passed in using the
|
||
|
// hostname argument. Generated urls will be of the form
|
||
|
// "<bucket-bound-hostname>/<object-name>". See
|
||
|
// https://cloud.google.com/storage/docs/request-endpoints#cname and
|
||
|
// https://cloud.google.com/load-balancing/docs/https/adding-backend-buckets-to-load-balancers
|
||
|
// for details. Note that for CNAMEs, only HTTP is supported, so Insecure must
|
||
|
// be set to true.
|
||
|
func BucketBoundHostname(hostname string) URLStyle {
|
||
|
return bucketBoundHostname{hostname: hostname}
|
||
|
}
|
||
|
|
||
|
// SignedURLOptions allows you to restrict the access to the signed URL.
|
||
|
type SignedURLOptions struct {
|
||
|
// GoogleAccessID represents the authorizer of the signed URL generation.
|
||
|
// It is typically the Google service account client email address from
|
||
|
// the Google Developers Console in the form of "xxx@developer.gserviceaccount.com".
|
||
|
// Required.
|
||
|
GoogleAccessID string
|
||
|
|
||
|
// PrivateKey is the Google service account private key. It is obtainable
|
||
|
// from the Google Developers Console.
|
||
|
// At https://console.developers.google.com/project/<your-project-id>/apiui/credential,
|
||
|
// create a service account client ID or reuse one of your existing service account
|
||
|
// credentials. Click on the "Generate new P12 key" to generate and download
|
||
|
// a new private key. Once you download the P12 file, use the following command
|
||
|
// to convert it into a PEM file.
|
||
|
//
|
||
|
// $ openssl pkcs12 -in key.p12 -passin pass:notasecret -out key.pem -nodes
|
||
|
//
|
||
|
// Provide the contents of the PEM file as a byte slice.
|
||
|
// Exactly one of PrivateKey or SignBytes must be non-nil.
|
||
|
PrivateKey []byte
|
||
|
|
||
|
// SignBytes is a function for implementing custom signing. For example, if
|
||
|
// your application is running on Google App Engine, you can use
|
||
|
// appengine's internal signing function:
|
||
|
// ctx := appengine.NewContext(request)
|
||
|
// acc, _ := appengine.ServiceAccount(ctx)
|
||
|
// url, err := SignedURL("bucket", "object", &SignedURLOptions{
|
||
|
// GoogleAccessID: acc,
|
||
|
// SignBytes: func(b []byte) ([]byte, error) {
|
||
|
// _, signedBytes, err := appengine.SignBytes(ctx, b)
|
||
|
// return signedBytes, err
|
||
|
// },
|
||
|
// // etc.
|
||
|
// })
|
||
|
//
|
||
|
// Exactly one of PrivateKey or SignBytes must be non-nil.
|
||
|
SignBytes func([]byte) ([]byte, error)
|
||
|
|
||
|
// Method is the HTTP method to be used with the signed URL.
|
||
|
// Signed URLs can be used with GET, HEAD, PUT, and DELETE requests.
|
||
|
// Required.
|
||
|
Method string
|
||
|
|
||
|
// Expires is the expiration time on the signed URL. It must be
|
||
|
// a datetime in the future. For SigningSchemeV4, the expiration may be no
|
||
|
// more than seven days in the future.
|
||
|
// Required.
|
||
|
Expires time.Time
|
||
|
|
||
|
// ContentType is the content type header the client must provide
|
||
|
// to use the generated signed URL.
|
||
|
// Optional.
|
||
|
ContentType string
|
||
|
|
||
|
// Headers is a list of extension headers the client must provide
|
||
|
// in order to use the generated signed URL. Each must be a string of the
|
||
|
// form "key:values", with multiple values separated by a semicolon.
|
||
|
// Optional.
|
||
|
Headers []string
|
||
|
|
||
|
// QueryParameters is a map of additional query parameters. When
|
||
|
// SigningScheme is V4, this is used in computing the signature, and the
|
||
|
// client must use the same query parameters when using the generated signed
|
||
|
// URL.
|
||
|
// Optional.
|
||
|
QueryParameters url.Values
|
||
|
|
||
|
// MD5 is the base64 encoded MD5 checksum of the file.
|
||
|
// If provided, the client should provide the exact value on the request
|
||
|
// header in order to use the signed URL.
|
||
|
// Optional.
|
||
|
MD5 string
|
||
|
|
||
|
// Style provides options for the type of URL to use. Options are
|
||
|
// PathStyle (default), BucketBoundHostname, and VirtualHostedStyle. See
|
||
|
// https://cloud.google.com/storage/docs/request-endpoints for details.
|
||
|
// Only supported for V4 signing.
|
||
|
// Optional.
|
||
|
Style URLStyle
|
||
|
|
||
|
// Insecure determines whether the signed URL should use HTTPS (default) or
|
||
|
// HTTP.
|
||
|
// Only supported for V4 signing.
|
||
|
// Optional.
|
||
|
Insecure bool
|
||
|
|
||
|
// Scheme determines the version of URL signing to use. Default is
|
||
|
// SigningSchemeV2.
|
||
|
Scheme SigningScheme
|
||
|
}
|
||
|
|
||
|
var (
|
||
|
tabRegex = regexp.MustCompile(`[\t]+`)
|
||
|
// I was tempted to call this spacex. :)
|
||
|
spaceRegex = regexp.MustCompile(` +`)
|
||
|
|
||
|
canonicalHeaderRegexp = regexp.MustCompile(`(?i)^(x-goog-[^:]+):(.*)?$`)
|
||
|
excludedCanonicalHeaders = map[string]bool{
|
||
|
"x-goog-encryption-key": true,
|
||
|
"x-goog-encryption-key-sha256": true,
|
||
|
}
|
||
|
)
|
||
|
|
||
|
// v2SanitizeHeaders applies the specifications for canonical extension headers at
|
||
|
// https://cloud.google.com/storage/docs/access-control/signed-urls-v2#about-canonical-extension-headers
|
||
|
func v2SanitizeHeaders(hdrs []string) []string {
|
||
|
headerMap := map[string][]string{}
|
||
|
for _, hdr := range hdrs {
|
||
|
// No leading or trailing whitespaces.
|
||
|
sanitizedHeader := strings.TrimSpace(hdr)
|
||
|
|
||
|
var header, value string
|
||
|
// Only keep canonical headers, discard any others.
|
||
|
headerMatches := canonicalHeaderRegexp.FindStringSubmatch(sanitizedHeader)
|
||
|
if len(headerMatches) == 0 {
|
||
|
continue
|
||
|
}
|
||
|
header = headerMatches[1]
|
||
|
value = headerMatches[2]
|
||
|
|
||
|
header = strings.ToLower(strings.TrimSpace(header))
|
||
|
value = strings.TrimSpace(value)
|
||
|
|
||
|
if excludedCanonicalHeaders[header] {
|
||
|
// Do not keep any deliberately excluded canonical headers when signing.
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
if len(value) > 0 {
|
||
|
// Remove duplicate headers by appending the values of duplicates
|
||
|
// in their order of appearance.
|
||
|
headerMap[header] = append(headerMap[header], value)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
var sanitizedHeaders []string
|
||
|
for header, values := range headerMap {
|
||
|
// There should be no spaces around the colon separating the header name
|
||
|
// from the header value or around the values themselves. The values
|
||
|
// should be separated by commas.
|
||
|
//
|
||
|
// NOTE: The semantics for headers without a value are not clear.
|
||
|
// However from specifications these should be edge-cases anyway and we
|
||
|
// should assume that there will be no canonical headers using empty
|
||
|
// values. Any such headers are discarded at the regexp stage above.
|
||
|
sanitizedHeaders = append(sanitizedHeaders, fmt.Sprintf("%s:%s", header, strings.Join(values, ",")))
|
||
|
}
|
||
|
sort.Strings(sanitizedHeaders)
|
||
|
return sanitizedHeaders
|
||
|
}
|
||
|
|
||
|
// v4SanitizeHeaders applies the specifications for canonical extension headers
|
||
|
// at https://cloud.google.com/storage/docs/authentication/canonical-requests#about-headers.
|
||
|
//
|
||
|
// V4 does a couple things differently from V2:
|
||
|
// - Headers get sorted by key, instead of by key:value. We do this in
|
||
|
// signedURLV4.
|
||
|
// - There's no canonical regexp: we simply split headers on :.
|
||
|
// - We don't exclude canonical headers.
|
||
|
// - We replace leading and trailing spaces in header values, like v2, but also
|
||
|
// all intermediate space duplicates get stripped. That is, there's only ever
|
||
|
// a single consecutive space.
|
||
|
func v4SanitizeHeaders(hdrs []string) []string {
|
||
|
headerMap := map[string][]string{}
|
||
|
for _, hdr := range hdrs {
|
||
|
// No leading or trailing whitespaces.
|
||
|
sanitizedHeader := strings.TrimSpace(hdr)
|
||
|
|
||
|
var key, value string
|
||
|
headerMatches := strings.Split(sanitizedHeader, ":")
|
||
|
if len(headerMatches) < 2 {
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
key = headerMatches[0]
|
||
|
value = headerMatches[1]
|
||
|
|
||
|
key = strings.ToLower(strings.TrimSpace(key))
|
||
|
value = strings.TrimSpace(value)
|
||
|
value = string(spaceRegex.ReplaceAll([]byte(value), []byte(" ")))
|
||
|
value = string(tabRegex.ReplaceAll([]byte(value), []byte("\t")))
|
||
|
|
||
|
if len(value) > 0 {
|
||
|
// Remove duplicate headers by appending the values of duplicates
|
||
|
// in their order of appearance.
|
||
|
headerMap[key] = append(headerMap[key], value)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
var sanitizedHeaders []string
|
||
|
for header, values := range headerMap {
|
||
|
// There should be no spaces around the colon separating the header name
|
||
|
// from the header value or around the values themselves. The values
|
||
|
// should be separated by commas.
|
||
|
//
|
||
|
// NOTE: The semantics for headers without a value are not clear.
|
||
|
// However from specifications these should be edge-cases anyway and we
|
||
|
// should assume that there will be no canonical headers using empty
|
||
|
// values. Any such headers are discarded at the regexp stage above.
|
||
|
sanitizedHeaders = append(sanitizedHeaders, fmt.Sprintf("%s:%s", header, strings.Join(values, ",")))
|
||
|
}
|
||
|
return sanitizedHeaders
|
||
|
}
|
||
|
|
||
|
// SignedURL returns a URL for the specified object. Signed URLs allow
|
||
|
// the users access to a restricted resource for a limited time without having a
|
||
|
// Google account or signing in. For more information about the signed
|
||
|
// URLs, see https://cloud.google.com/storage/docs/accesscontrol#Signed-URLs.
|
||
|
func SignedURL(bucket, name string, opts *SignedURLOptions) (string, error) {
|
||
|
now := utcNow()
|
||
|
if err := validateOptions(opts, now); err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
|
||
|
switch opts.Scheme {
|
||
|
case SigningSchemeV2:
|
||
|
opts.Headers = v2SanitizeHeaders(opts.Headers)
|
||
|
return signedURLV2(bucket, name, opts)
|
||
|
case SigningSchemeV4:
|
||
|
opts.Headers = v4SanitizeHeaders(opts.Headers)
|
||
|
return signedURLV4(bucket, name, opts, now)
|
||
|
default: // SigningSchemeDefault
|
||
|
opts.Headers = v2SanitizeHeaders(opts.Headers)
|
||
|
return signedURLV2(bucket, name, opts)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func validateOptions(opts *SignedURLOptions, now time.Time) error {
|
||
|
if opts == nil {
|
||
|
return errors.New("storage: missing required SignedURLOptions")
|
||
|
}
|
||
|
if opts.GoogleAccessID == "" {
|
||
|
return errors.New("storage: missing required GoogleAccessID")
|
||
|
}
|
||
|
if (opts.PrivateKey == nil) == (opts.SignBytes == nil) {
|
||
|
return errors.New("storage: exactly one of PrivateKey or SignedBytes must be set")
|
||
|
}
|
||
|
opts.Method = strings.ToUpper(opts.Method)
|
||
|
if _, ok := signedURLMethods[opts.Method]; !ok {
|
||
|
return errMethodNotValid
|
||
|
}
|
||
|
if opts.Expires.IsZero() {
|
||
|
return errors.New("storage: missing required expires option")
|
||
|
}
|
||
|
if opts.MD5 != "" {
|
||
|
md5, err := base64.StdEncoding.DecodeString(opts.MD5)
|
||
|
if err != nil || len(md5) != 16 {
|
||
|
return errors.New("storage: invalid MD5 checksum")
|
||
|
}
|
||
|
}
|
||
|
if opts.Style == nil {
|
||
|
opts.Style = PathStyle()
|
||
|
}
|
||
|
if _, ok := opts.Style.(pathStyle); !ok && opts.Scheme == SigningSchemeV2 {
|
||
|
return errors.New("storage: only path-style URLs are permitted with SigningSchemeV2")
|
||
|
}
|
||
|
if opts.Scheme == SigningSchemeV4 {
|
||
|
cutoff := now.Add(604801 * time.Second) // 7 days + 1 second
|
||
|
if !opts.Expires.Before(cutoff) {
|
||
|
return errors.New("storage: expires must be within seven days from now")
|
||
|
}
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
const (
|
||
|
iso8601 = "20060102T150405Z"
|
||
|
yearMonthDay = "20060102"
|
||
|
)
|
||
|
|
||
|
// utcNow returns the current time in UTC and is a variable to allow for
|
||
|
// reassignment in tests to provide deterministic signed URL values.
|
||
|
var utcNow = func() time.Time {
|
||
|
return time.Now().UTC()
|
||
|
}
|
||
|
|
||
|
// extractHeaderNames takes in a series of key:value headers and returns the
|
||
|
// header names only.
|
||
|
func extractHeaderNames(kvs []string) []string {
|
||
|
var res []string
|
||
|
for _, header := range kvs {
|
||
|
nameValue := strings.Split(header, ":")
|
||
|
res = append(res, nameValue[0])
|
||
|
}
|
||
|
return res
|
||
|
}
|
||
|
|
||
|
// pathEncodeV4 creates an encoded string that matches the v4 signature spec.
|
||
|
// Following the spec precisely is necessary in order to ensure that the URL
|
||
|
// and signing string are correctly formed, and Go's url.PathEncode and
|
||
|
// url.QueryEncode don't generate an exact match without some additional logic.
|
||
|
func pathEncodeV4(path string) string {
|
||
|
segments := strings.Split(path, "/")
|
||
|
var encodedSegments []string
|
||
|
for _, s := range segments {
|
||
|
encodedSegments = append(encodedSegments, url.QueryEscape(s))
|
||
|
}
|
||
|
encodedStr := strings.Join(encodedSegments, "/")
|
||
|
encodedStr = strings.Replace(encodedStr, "+", "%20", -1)
|
||
|
return encodedStr
|
||
|
}
|
||
|
|
||
|
// signedURLV4 creates a signed URL using the sigV4 algorithm.
|
||
|
func signedURLV4(bucket, name string, opts *SignedURLOptions, now time.Time) (string, error) {
|
||
|
buf := &bytes.Buffer{}
|
||
|
fmt.Fprintf(buf, "%s\n", opts.Method)
|
||
|
|
||
|
u := &url.URL{Path: opts.Style.path(bucket, name)}
|
||
|
u.RawPath = pathEncodeV4(u.Path)
|
||
|
|
||
|
// Note: we have to add a / here because GCS does so auto-magically, despite
|
||
|
// our encoding not doing so (and we have to exactly match their
|
||
|
// canonical query).
|
||
|
fmt.Fprintf(buf, "/%s\n", u.RawPath)
|
||
|
|
||
|
headerNames := append(extractHeaderNames(opts.Headers), "host")
|
||
|
if opts.ContentType != "" {
|
||
|
headerNames = append(headerNames, "content-type")
|
||
|
}
|
||
|
if opts.MD5 != "" {
|
||
|
headerNames = append(headerNames, "content-md5")
|
||
|
}
|
||
|
sort.Strings(headerNames)
|
||
|
signedHeaders := strings.Join(headerNames, ";")
|
||
|
timestamp := now.Format(iso8601)
|
||
|
credentialScope := fmt.Sprintf("%s/auto/storage/goog4_request", now.Format(yearMonthDay))
|
||
|
canonicalQueryString := url.Values{
|
||
|
"X-Goog-Algorithm": {"GOOG4-RSA-SHA256"},
|
||
|
"X-Goog-Credential": {fmt.Sprintf("%s/%s", opts.GoogleAccessID, credentialScope)},
|
||
|
"X-Goog-Date": {timestamp},
|
||
|
"X-Goog-Expires": {fmt.Sprintf("%d", int(opts.Expires.Sub(now).Seconds()))},
|
||
|
"X-Goog-SignedHeaders": {signedHeaders},
|
||
|
}
|
||
|
// Add user-supplied query parameters to the canonical query string. For V4,
|
||
|
// it's necessary to include these.
|
||
|
for k, v := range opts.QueryParameters {
|
||
|
canonicalQueryString[k] = append(canonicalQueryString[k], v...)
|
||
|
}
|
||
|
// url.Values.Encode escaping is correct, except that a space must be replaced
|
||
|
// by `%20` rather than `+`.
|
||
|
escapedQuery := strings.Replace(canonicalQueryString.Encode(), "+", "%20", -1)
|
||
|
fmt.Fprintf(buf, "%s\n", escapedQuery)
|
||
|
|
||
|
// Fill in the hostname based on the desired URL style.
|
||
|
u.Host = opts.Style.host(bucket)
|
||
|
|
||
|
// Fill in the URL scheme.
|
||
|
if opts.Insecure {
|
||
|
u.Scheme = "http"
|
||
|
} else {
|
||
|
u.Scheme = "https"
|
||
|
}
|
||
|
|
||
|
var headersWithValue []string
|
||
|
headersWithValue = append(headersWithValue, "host:"+u.Host)
|
||
|
headersWithValue = append(headersWithValue, opts.Headers...)
|
||
|
if opts.ContentType != "" {
|
||
|
headersWithValue = append(headersWithValue, "content-type:"+opts.ContentType)
|
||
|
}
|
||
|
if opts.MD5 != "" {
|
||
|
headersWithValue = append(headersWithValue, "content-md5:"+opts.MD5)
|
||
|
}
|
||
|
// Trim extra whitespace from headers and replace with a single space.
|
||
|
var trimmedHeaders []string
|
||
|
for _, h := range headersWithValue {
|
||
|
trimmedHeaders = append(trimmedHeaders, strings.Join(strings.Fields(h), " "))
|
||
|
}
|
||
|
canonicalHeaders := strings.Join(sortHeadersByKey(trimmedHeaders), "\n")
|
||
|
fmt.Fprintf(buf, "%s\n\n", canonicalHeaders)
|
||
|
fmt.Fprintf(buf, "%s\n", signedHeaders)
|
||
|
|
||
|
// If the user provides a value for X-Goog-Content-SHA256, we must use
|
||
|
// that value in the request string. If not, we use UNSIGNED-PAYLOAD.
|
||
|
sha256Header := false
|
||
|
for _, h := range trimmedHeaders {
|
||
|
if strings.HasPrefix(strings.ToLower(h), "x-goog-content-sha256") && strings.Contains(h, ":") {
|
||
|
sha256Header = true
|
||
|
fmt.Fprintf(buf, "%s", strings.SplitN(h, ":", 2)[1])
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
if !sha256Header {
|
||
|
fmt.Fprint(buf, "UNSIGNED-PAYLOAD")
|
||
|
}
|
||
|
|
||
|
sum := sha256.Sum256(buf.Bytes())
|
||
|
hexDigest := hex.EncodeToString(sum[:])
|
||
|
signBuf := &bytes.Buffer{}
|
||
|
fmt.Fprint(signBuf, "GOOG4-RSA-SHA256\n")
|
||
|
fmt.Fprintf(signBuf, "%s\n", timestamp)
|
||
|
fmt.Fprintf(signBuf, "%s\n", credentialScope)
|
||
|
fmt.Fprintf(signBuf, "%s", hexDigest)
|
||
|
|
||
|
signBytes := opts.SignBytes
|
||
|
if opts.PrivateKey != nil {
|
||
|
key, err := parseKey(opts.PrivateKey)
|
||
|
if err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
signBytes = func(b []byte) ([]byte, error) {
|
||
|
sum := sha256.Sum256(b)
|
||
|
return rsa.SignPKCS1v15(
|
||
|
rand.Reader,
|
||
|
key,
|
||
|
crypto.SHA256,
|
||
|
sum[:],
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
b, err := signBytes(signBuf.Bytes())
|
||
|
if err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
signature := hex.EncodeToString(b)
|
||
|
canonicalQueryString.Set("X-Goog-Signature", string(signature))
|
||
|
u.RawQuery = canonicalQueryString.Encode()
|
||
|
return u.String(), nil
|
||
|
}
|
||
|
|
||
|
// takes a list of headerKey:headervalue1,headervalue2,etc and sorts by header
|
||
|
// key.
|
||
|
func sortHeadersByKey(hdrs []string) []string {
|
||
|
headersMap := map[string]string{}
|
||
|
var headersKeys []string
|
||
|
for _, h := range hdrs {
|
||
|
parts := strings.Split(h, ":")
|
||
|
k := parts[0]
|
||
|
v := parts[1]
|
||
|
headersMap[k] = v
|
||
|
headersKeys = append(headersKeys, k)
|
||
|
}
|
||
|
sort.Strings(headersKeys)
|
||
|
var sorted []string
|
||
|
for _, k := range headersKeys {
|
||
|
v := headersMap[k]
|
||
|
sorted = append(sorted, fmt.Sprintf("%s:%s", k, v))
|
||
|
}
|
||
|
return sorted
|
||
|
}
|
||
|
|
||
|
func signedURLV2(bucket, name string, opts *SignedURLOptions) (string, error) {
|
||
|
signBytes := opts.SignBytes
|
||
|
if opts.PrivateKey != nil {
|
||
|
key, err := parseKey(opts.PrivateKey)
|
||
|
if err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
signBytes = func(b []byte) ([]byte, error) {
|
||
|
sum := sha256.Sum256(b)
|
||
|
return rsa.SignPKCS1v15(
|
||
|
rand.Reader,
|
||
|
key,
|
||
|
crypto.SHA256,
|
||
|
sum[:],
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
u := &url.URL{
|
||
|
Path: fmt.Sprintf("/%s/%s", bucket, name),
|
||
|
}
|
||
|
|
||
|
buf := &bytes.Buffer{}
|
||
|
fmt.Fprintf(buf, "%s\n", opts.Method)
|
||
|
fmt.Fprintf(buf, "%s\n", opts.MD5)
|
||
|
fmt.Fprintf(buf, "%s\n", opts.ContentType)
|
||
|
fmt.Fprintf(buf, "%d\n", opts.Expires.Unix())
|
||
|
if len(opts.Headers) > 0 {
|
||
|
fmt.Fprintf(buf, "%s\n", strings.Join(opts.Headers, "\n"))
|
||
|
}
|
||
|
fmt.Fprintf(buf, "%s", u.String())
|
||
|
|
||
|
b, err := signBytes(buf.Bytes())
|
||
|
if err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
encoded := base64.StdEncoding.EncodeToString(b)
|
||
|
u.Scheme = "https"
|
||
|
u.Host = "storage.googleapis.com"
|
||
|
q := u.Query()
|
||
|
q.Set("GoogleAccessId", opts.GoogleAccessID)
|
||
|
q.Set("Expires", fmt.Sprintf("%d", opts.Expires.Unix()))
|
||
|
q.Set("Signature", string(encoded))
|
||
|
u.RawQuery = q.Encode()
|
||
|
return u.String(), nil
|
||
|
}
|
||
|
|
||
|
// ObjectHandle provides operations on an object in a Google Cloud Storage bucket.
|
||
|
// Use BucketHandle.Object to get a handle.
|
||
|
type ObjectHandle struct {
|
||
|
c *Client
|
||
|
bucket string
|
||
|
object string
|
||
|
acl ACLHandle
|
||
|
gen int64 // a negative value indicates latest
|
||
|
conds *Conditions
|
||
|
encryptionKey []byte // AES-256 key
|
||
|
userProject string // for requester-pays buckets
|
||
|
readCompressed bool // Accept-Encoding: gzip
|
||
|
}
|
||
|
|
||
|
// ACL provides access to the object's access control list.
|
||
|
// This controls who can read and write this object.
|
||
|
// This call does not perform any network operations.
|
||
|
func (o *ObjectHandle) ACL() *ACLHandle {
|
||
|
return &o.acl
|
||
|
}
|
||
|
|
||
|
// Generation returns a new ObjectHandle that operates on a specific generation
|
||
|
// of the object.
|
||
|
// By default, the handle operates on the latest generation. Not
|
||
|
// all operations work when given a specific generation; check the API
|
||
|
// endpoints at https://cloud.google.com/storage/docs/json_api/ for details.
|
||
|
func (o *ObjectHandle) Generation(gen int64) *ObjectHandle {
|
||
|
o2 := *o
|
||
|
o2.gen = gen
|
||
|
return &o2
|
||
|
}
|
||
|
|
||
|
// If returns a new ObjectHandle that applies a set of preconditions.
|
||
|
// Preconditions already set on the ObjectHandle are ignored.
|
||
|
// Operations on the new handle will return an error if the preconditions are not
|
||
|
// satisfied. See https://cloud.google.com/storage/docs/generations-preconditions
|
||
|
// for more details.
|
||
|
func (o *ObjectHandle) If(conds Conditions) *ObjectHandle {
|
||
|
o2 := *o
|
||
|
o2.conds = &conds
|
||
|
return &o2
|
||
|
}
|
||
|
|
||
|
// Key returns a new ObjectHandle that uses the supplied encryption
|
||
|
// key to encrypt and decrypt the object's contents.
|
||
|
//
|
||
|
// Encryption key must be a 32-byte AES-256 key.
|
||
|
// See https://cloud.google.com/storage/docs/encryption for details.
|
||
|
func (o *ObjectHandle) Key(encryptionKey []byte) *ObjectHandle {
|
||
|
o2 := *o
|
||
|
o2.encryptionKey = encryptionKey
|
||
|
return &o2
|
||
|
}
|
||
|
|
||
|
// Attrs returns meta information about the object.
|
||
|
// ErrObjectNotExist will be returned if the object is not found.
|
||
|
func (o *ObjectHandle) Attrs(ctx context.Context) (attrs *ObjectAttrs, err error) {
|
||
|
ctx = trace.StartSpan(ctx, "cloud.google.com/go/storage.Object.Attrs")
|
||
|
defer func() { trace.EndSpan(ctx, err) }()
|
||
|
|
||
|
if err := o.validate(); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
call := o.c.raw.Objects.Get(o.bucket, o.object).Projection("full").Context(ctx)
|
||
|
if err := applyConds("Attrs", o.gen, o.conds, call); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
if o.userProject != "" {
|
||
|
call.UserProject(o.userProject)
|
||
|
}
|
||
|
if err := setEncryptionHeaders(call.Header(), o.encryptionKey, false); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
var obj *raw.Object
|
||
|
setClientHeader(call.Header())
|
||
|
err = runWithRetry(ctx, func() error { obj, err = call.Do(); return err })
|
||
|
if e, ok := err.(*googleapi.Error); ok && e.Code == http.StatusNotFound {
|
||
|
return nil, ErrObjectNotExist
|
||
|
}
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
return newObject(obj), nil
|
||
|
}
|
||
|
|
||
|
// Update updates an object with the provided attributes.
|
||
|
// All zero-value attributes are ignored.
|
||
|
// ErrObjectNotExist will be returned if the object is not found.
|
||
|
func (o *ObjectHandle) Update(ctx context.Context, uattrs ObjectAttrsToUpdate) (oa *ObjectAttrs, err error) {
|
||
|
ctx = trace.StartSpan(ctx, "cloud.google.com/go/storage.Object.Update")
|
||
|
defer func() { trace.EndSpan(ctx, err) }()
|
||
|
|
||
|
if err := o.validate(); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
var attrs ObjectAttrs
|
||
|
// Lists of fields to send, and set to null, in the JSON.
|
||
|
var forceSendFields, nullFields []string
|
||
|
if uattrs.ContentType != nil {
|
||
|
attrs.ContentType = optional.ToString(uattrs.ContentType)
|
||
|
// For ContentType, sending the empty string is a no-op.
|
||
|
// Instead we send a null.
|
||
|
if attrs.ContentType == "" {
|
||
|
nullFields = append(nullFields, "ContentType")
|
||
|
} else {
|
||
|
forceSendFields = append(forceSendFields, "ContentType")
|
||
|
}
|
||
|
}
|
||
|
if uattrs.ContentLanguage != nil {
|
||
|
attrs.ContentLanguage = optional.ToString(uattrs.ContentLanguage)
|
||
|
// For ContentLanguage it's an error to send the empty string.
|
||
|
// Instead we send a null.
|
||
|
if attrs.ContentLanguage == "" {
|
||
|
nullFields = append(nullFields, "ContentLanguage")
|
||
|
} else {
|
||
|
forceSendFields = append(forceSendFields, "ContentLanguage")
|
||
|
}
|
||
|
}
|
||
|
if uattrs.ContentEncoding != nil {
|
||
|
attrs.ContentEncoding = optional.ToString(uattrs.ContentEncoding)
|
||
|
forceSendFields = append(forceSendFields, "ContentEncoding")
|
||
|
}
|
||
|
if uattrs.ContentDisposition != nil {
|
||
|
attrs.ContentDisposition = optional.ToString(uattrs.ContentDisposition)
|
||
|
forceSendFields = append(forceSendFields, "ContentDisposition")
|
||
|
}
|
||
|
if uattrs.CacheControl != nil {
|
||
|
attrs.CacheControl = optional.ToString(uattrs.CacheControl)
|
||
|
forceSendFields = append(forceSendFields, "CacheControl")
|
||
|
}
|
||
|
if uattrs.EventBasedHold != nil {
|
||
|
attrs.EventBasedHold = optional.ToBool(uattrs.EventBasedHold)
|
||
|
forceSendFields = append(forceSendFields, "EventBasedHold")
|
||
|
}
|
||
|
if uattrs.TemporaryHold != nil {
|
||
|
attrs.TemporaryHold = optional.ToBool(uattrs.TemporaryHold)
|
||
|
forceSendFields = append(forceSendFields, "TemporaryHold")
|
||
|
}
|
||
|
if !uattrs.CustomTime.IsZero() {
|
||
|
attrs.CustomTime = uattrs.CustomTime
|
||
|
forceSendFields = append(forceSendFields, "CustomTime")
|
||
|
}
|
||
|
if uattrs.Metadata != nil {
|
||
|
attrs.Metadata = uattrs.Metadata
|
||
|
if len(attrs.Metadata) == 0 {
|
||
|
// Sending the empty map is a no-op. We send null instead.
|
||
|
nullFields = append(nullFields, "Metadata")
|
||
|
} else {
|
||
|
forceSendFields = append(forceSendFields, "Metadata")
|
||
|
}
|
||
|
}
|
||
|
if uattrs.ACL != nil {
|
||
|
attrs.ACL = uattrs.ACL
|
||
|
// It's an error to attempt to delete the ACL, so
|
||
|
// we don't append to nullFields here.
|
||
|
forceSendFields = append(forceSendFields, "Acl")
|
||
|
}
|
||
|
rawObj := attrs.toRawObject(o.bucket)
|
||
|
rawObj.ForceSendFields = forceSendFields
|
||
|
rawObj.NullFields = nullFields
|
||
|
call := o.c.raw.Objects.Patch(o.bucket, o.object, rawObj).Projection("full").Context(ctx)
|
||
|
if err := applyConds("Update", o.gen, o.conds, call); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
if o.userProject != "" {
|
||
|
call.UserProject(o.userProject)
|
||
|
}
|
||
|
if uattrs.PredefinedACL != "" {
|
||
|
call.PredefinedAcl(uattrs.PredefinedACL)
|
||
|
}
|
||
|
if err := setEncryptionHeaders(call.Header(), o.encryptionKey, false); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
var obj *raw.Object
|
||
|
setClientHeader(call.Header())
|
||
|
err = runWithRetry(ctx, func() error { obj, err = call.Do(); return err })
|
||
|
if e, ok := err.(*googleapi.Error); ok && e.Code == http.StatusNotFound {
|
||
|
return nil, ErrObjectNotExist
|
||
|
}
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
return newObject(obj), nil
|
||
|
}
|
||
|
|
||
|
// BucketName returns the name of the bucket.
|
||
|
func (o *ObjectHandle) BucketName() string {
|
||
|
return o.bucket
|
||
|
}
|
||
|
|
||
|
// ObjectName returns the name of the object.
|
||
|
func (o *ObjectHandle) ObjectName() string {
|
||
|
return o.object
|
||
|
}
|
||
|
|
||
|
// ObjectAttrsToUpdate is used to update the attributes of an object.
|
||
|
// Only fields set to non-nil values will be updated.
|
||
|
// Set a field to its zero value to delete it.
|
||
|
//
|
||
|
// For example, to change ContentType and delete ContentEncoding and
|
||
|
// Metadata, use
|
||
|
// ObjectAttrsToUpdate{
|
||
|
// ContentType: "text/html",
|
||
|
// ContentEncoding: "",
|
||
|
// Metadata: map[string]string{},
|
||
|
// }
|
||
|
type ObjectAttrsToUpdate struct {
|
||
|
EventBasedHold optional.Bool
|
||
|
TemporaryHold optional.Bool
|
||
|
ContentType optional.String
|
||
|
ContentLanguage optional.String
|
||
|
ContentEncoding optional.String
|
||
|
ContentDisposition optional.String
|
||
|
CacheControl optional.String
|
||
|
CustomTime time.Time
|
||
|
Metadata map[string]string // set to map[string]string{} to delete
|
||
|
ACL []ACLRule
|
||
|
|
||
|
// If not empty, applies a predefined set of access controls. ACL must be nil.
|
||
|
// See https://cloud.google.com/storage/docs/json_api/v1/objects/patch.
|
||
|
PredefinedACL string
|
||
|
}
|
||
|
|
||
|
// Delete deletes the single specified object.
|
||
|
func (o *ObjectHandle) Delete(ctx context.Context) error {
|
||
|
if err := o.validate(); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
call := o.c.raw.Objects.Delete(o.bucket, o.object).Context(ctx)
|
||
|
if err := applyConds("Delete", o.gen, o.conds, call); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
if o.userProject != "" {
|
||
|
call.UserProject(o.userProject)
|
||
|
}
|
||
|
// Encryption doesn't apply to Delete.
|
||
|
setClientHeader(call.Header())
|
||
|
err := runWithRetry(ctx, func() error { return call.Do() })
|
||
|
switch e := err.(type) {
|
||
|
case nil:
|
||
|
return nil
|
||
|
case *googleapi.Error:
|
||
|
if e.Code == http.StatusNotFound {
|
||
|
return ErrObjectNotExist
|
||
|
}
|
||
|
}
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
// ReadCompressed when true causes the read to happen without decompressing.
|
||
|
func (o *ObjectHandle) ReadCompressed(compressed bool) *ObjectHandle {
|
||
|
o2 := *o
|
||
|
o2.readCompressed = compressed
|
||
|
return &o2
|
||
|
}
|
||
|
|
||
|
// NewWriter returns a storage Writer that writes to the GCS object
|
||
|
// associated with this ObjectHandle.
|
||
|
//
|
||
|
// A new object will be created unless an object with this name already exists.
|
||
|
// Otherwise any previous object with the same name will be replaced.
|
||
|
// The object will not be available (and any previous object will remain)
|
||
|
// until Close has been called.
|
||
|
//
|
||
|
// Attributes can be set on the object by modifying the returned Writer's
|
||
|
// ObjectAttrs field before the first call to Write. If no ContentType
|
||
|
// attribute is specified, the content type will be automatically sniffed
|
||
|
// using net/http.DetectContentType.
|
||
|
//
|
||
|
// It is the caller's responsibility to call Close when writing is done. To
|
||
|
// stop writing without saving the data, cancel the context.
|
||
|
func (o *ObjectHandle) NewWriter(ctx context.Context) *Writer {
|
||
|
return &Writer{
|
||
|
ctx: ctx,
|
||
|
o: o,
|
||
|
donec: make(chan struct{}),
|
||
|
ObjectAttrs: ObjectAttrs{Name: o.object},
|
||
|
ChunkSize: googleapi.DefaultUploadChunkSize,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (o *ObjectHandle) validate() error {
|
||
|
if o.bucket == "" {
|
||
|
return errors.New("storage: bucket name is empty")
|
||
|
}
|
||
|
if o.object == "" {
|
||
|
return errors.New("storage: object name is empty")
|
||
|
}
|
||
|
if !utf8.ValidString(o.object) {
|
||
|
return fmt.Errorf("storage: object name %q is not valid UTF-8", o.object)
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// parseKey converts the binary contents of a private key file to an
|
||
|
// *rsa.PrivateKey. It detects whether the private key is in a PEM container or
|
||
|
// not. If so, it extracts the private key from PEM container before
|
||
|
// conversion. It only supports PEM containers with no passphrase.
|
||
|
func parseKey(key []byte) (*rsa.PrivateKey, error) {
|
||
|
if block, _ := pem.Decode(key); block != nil {
|
||
|
key = block.Bytes
|
||
|
}
|
||
|
parsedKey, err := x509.ParsePKCS8PrivateKey(key)
|
||
|
if err != nil {
|
||
|
parsedKey, err = x509.ParsePKCS1PrivateKey(key)
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
}
|
||
|
parsed, ok := parsedKey.(*rsa.PrivateKey)
|
||
|
if !ok {
|
||
|
return nil, errors.New("oauth2: private key is invalid")
|
||
|
}
|
||
|
return parsed, nil
|
||
|
}
|
||
|
|
||
|
// toRawObject copies the editable attributes from o to the raw library's Object type.
|
||
|
func (o *ObjectAttrs) toRawObject(bucket string) *raw.Object {
|
||
|
var ret string
|
||
|
if !o.RetentionExpirationTime.IsZero() {
|
||
|
ret = o.RetentionExpirationTime.Format(time.RFC3339)
|
||
|
}
|
||
|
var ct string
|
||
|
if !o.CustomTime.IsZero() {
|
||
|
ct = o.CustomTime.Format(time.RFC3339)
|
||
|
}
|
||
|
return &raw.Object{
|
||
|
Bucket: bucket,
|
||
|
Name: o.Name,
|
||
|
EventBasedHold: o.EventBasedHold,
|
||
|
TemporaryHold: o.TemporaryHold,
|
||
|
RetentionExpirationTime: ret,
|
||
|
ContentType: o.ContentType,
|
||
|
ContentEncoding: o.ContentEncoding,
|
||
|
ContentLanguage: o.ContentLanguage,
|
||
|
CacheControl: o.CacheControl,
|
||
|
ContentDisposition: o.ContentDisposition,
|
||
|
StorageClass: o.StorageClass,
|
||
|
Acl: toRawObjectACL(o.ACL),
|
||
|
Metadata: o.Metadata,
|
||
|
CustomTime: ct,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// ObjectAttrs represents the metadata for a Google Cloud Storage (GCS) object.
|
||
|
type ObjectAttrs struct {
|
||
|
// Bucket is the name of the bucket containing this GCS object.
|
||
|
// This field is read-only.
|
||
|
Bucket string
|
||
|
|
||
|
// Name is the name of the object within the bucket.
|
||
|
// This field is read-only.
|
||
|
Name string
|
||
|
|
||
|
// ContentType is the MIME type of the object's content.
|
||
|
ContentType string
|
||
|
|
||
|
// ContentLanguage is the content language of the object's content.
|
||
|
ContentLanguage string
|
||
|
|
||
|
// CacheControl is the Cache-Control header to be sent in the response
|
||
|
// headers when serving the object data.
|
||
|
CacheControl string
|
||
|
|
||
|
// EventBasedHold specifies whether an object is under event-based hold. New
|
||
|
// objects created in a bucket whose DefaultEventBasedHold is set will
|
||
|
// default to that value.
|
||
|
EventBasedHold bool
|
||
|
|
||
|
// TemporaryHold specifies whether an object is under temporary hold. While
|
||
|
// this flag is set to true, the object is protected against deletion and
|
||
|
// overwrites.
|
||
|
TemporaryHold bool
|
||
|
|
||
|
// RetentionExpirationTime is a server-determined value that specifies the
|
||
|
// earliest time that the object's retention period expires.
|
||
|
// This is a read-only field.
|
||
|
RetentionExpirationTime time.Time
|
||
|
|
||
|
// ACL is the list of access control rules for the object.
|
||
|
ACL []ACLRule
|
||
|
|
||
|
// If not empty, applies a predefined set of access controls. It should be set
|
||
|
// only when writing, copying or composing an object. When copying or composing,
|
||
|
// it acts as the destinationPredefinedAcl parameter.
|
||
|
// PredefinedACL is always empty for ObjectAttrs returned from the service.
|
||
|
// See https://cloud.google.com/storage/docs/json_api/v1/objects/insert
|
||
|
// for valid values.
|
||
|
PredefinedACL string
|
||
|
|
||
|
// Owner is the owner of the object. This field is read-only.
|
||
|
//
|
||
|
// If non-zero, it is in the form of "user-<userId>".
|
||
|
Owner string
|
||
|
|
||
|
// Size is the length of the object's content. This field is read-only.
|
||
|
Size int64
|
||
|
|
||
|
// ContentEncoding is the encoding of the object's content.
|
||
|
ContentEncoding string
|
||
|
|
||
|
// ContentDisposition is the optional Content-Disposition header of the object
|
||
|
// sent in the response headers.
|
||
|
ContentDisposition string
|
||
|
|
||
|
// MD5 is the MD5 hash of the object's content. This field is read-only,
|
||
|
// except when used from a Writer. If set on a Writer, the uploaded
|
||
|
// data is rejected if its MD5 hash does not match this field.
|
||
|
MD5 []byte
|
||
|
|
||
|
// CRC32C is the CRC32 checksum of the object's content using the Castagnoli93
|
||
|
// polynomial. This field is read-only, except when used from a Writer or
|
||
|
// Composer. In those cases, if the SendCRC32C field in the Writer or Composer
|
||
|
// is set to is true, the uploaded data is rejected if its CRC32C hash does
|
||
|
// not match this field.
|
||
|
CRC32C uint32
|
||
|
|
||
|
// MediaLink is an URL to the object's content. This field is read-only.
|
||
|
MediaLink string
|
||
|
|
||
|
// Metadata represents user-provided metadata, in key/value pairs.
|
||
|
// It can be nil if no metadata is provided.
|
||
|
Metadata map[string]string
|
||
|
|
||
|
// Generation is the generation number of the object's content.
|
||
|
// This field is read-only.
|
||
|
Generation int64
|
||
|
|
||
|
// Metageneration is the version of the metadata for this
|
||
|
// object at this generation. This field is used for preconditions
|
||
|
// and for detecting changes in metadata. A metageneration number
|
||
|
// is only meaningful in the context of a particular generation
|
||
|
// of a particular object. This field is read-only.
|
||
|
Metageneration int64
|
||
|
|
||
|
// StorageClass is the storage class of the object. This defines
|
||
|
// how objects are stored and determines the SLA and the cost of storage.
|
||
|
// Typical values are "STANDARD", "NEARLINE", "COLDLINE" and "ARCHIVE".
|
||
|
// Defaults to "STANDARD".
|
||
|
// See https://cloud.google.com/storage/docs/storage-classes for all
|
||
|
// valid values.
|
||
|
StorageClass string
|
||
|
|
||
|
// Created is the time the object was created. This field is read-only.
|
||
|
Created time.Time
|
||
|
|
||
|
// Deleted is the time the object was deleted.
|
||
|
// If not deleted, it is the zero value. This field is read-only.
|
||
|
Deleted time.Time
|
||
|
|
||
|
// Updated is the creation or modification time of the object.
|
||
|
// For buckets with versioning enabled, changing an object's
|
||
|
// metadata does not change this property. This field is read-only.
|
||
|
Updated time.Time
|
||
|
|
||
|
// CustomerKeySHA256 is the base64-encoded SHA-256 hash of the
|
||
|
// customer-supplied encryption key for the object. It is empty if there is
|
||
|
// no customer-supplied encryption key.
|
||
|
// See // https://cloud.google.com/storage/docs/encryption for more about
|
||
|
// encryption in Google Cloud Storage.
|
||
|
CustomerKeySHA256 string
|
||
|
|
||
|
// Cloud KMS key name, in the form
|
||
|
// projects/P/locations/L/keyRings/R/cryptoKeys/K, used to encrypt this object,
|
||
|
// if the object is encrypted by such a key.
|
||
|
//
|
||
|
// Providing both a KMSKeyName and a customer-supplied encryption key (via
|
||
|
// ObjectHandle.Key) will result in an error when writing an object.
|
||
|
KMSKeyName string
|
||
|
|
||
|
// Prefix is set only for ObjectAttrs which represent synthetic "directory
|
||
|
// entries" when iterating over buckets using Query.Delimiter. See
|
||
|
// ObjectIterator.Next. When set, no other fields in ObjectAttrs will be
|
||
|
// populated.
|
||
|
Prefix string
|
||
|
|
||
|
// Etag is the HTTP/1.1 Entity tag for the object.
|
||
|
// This field is read-only.
|
||
|
Etag string
|
||
|
|
||
|
// A user-specified timestamp which can be applied to an object. This is
|
||
|
// typically set in order to use the CustomTimeBefore and DaysSinceCustomTime
|
||
|
// LifecycleConditions to manage object lifecycles.
|
||
|
//
|
||
|
// CustomTime cannot be removed once set on an object. It can be updated to a
|
||
|
// later value but not to an earlier one.
|
||
|
CustomTime time.Time
|
||
|
}
|
||
|
|
||
|
// convertTime converts a time in RFC3339 format to time.Time.
|
||
|
// If any error occurs in parsing, the zero-value time.Time is silently returned.
|
||
|
func convertTime(t string) time.Time {
|
||
|
var r time.Time
|
||
|
if t != "" {
|
||
|
r, _ = time.Parse(time.RFC3339, t)
|
||
|
}
|
||
|
return r
|
||
|
}
|
||
|
|
||
|
func newObject(o *raw.Object) *ObjectAttrs {
|
||
|
if o == nil {
|
||
|
return nil
|
||
|
}
|
||
|
owner := ""
|
||
|
if o.Owner != nil {
|
||
|
owner = o.Owner.Entity
|
||
|
}
|
||
|
md5, _ := base64.StdEncoding.DecodeString(o.Md5Hash)
|
||
|
crc32c, _ := decodeUint32(o.Crc32c)
|
||
|
var sha256 string
|
||
|
if o.CustomerEncryption != nil {
|
||
|
sha256 = o.CustomerEncryption.KeySha256
|
||
|
}
|
||
|
return &ObjectAttrs{
|
||
|
Bucket: o.Bucket,
|
||
|
Name: o.Name,
|
||
|
ContentType: o.ContentType,
|
||
|
ContentLanguage: o.ContentLanguage,
|
||
|
CacheControl: o.CacheControl,
|
||
|
EventBasedHold: o.EventBasedHold,
|
||
|
TemporaryHold: o.TemporaryHold,
|
||
|
RetentionExpirationTime: convertTime(o.RetentionExpirationTime),
|
||
|
ACL: toObjectACLRules(o.Acl),
|
||
|
Owner: owner,
|
||
|
ContentEncoding: o.ContentEncoding,
|
||
|
ContentDisposition: o.ContentDisposition,
|
||
|
Size: int64(o.Size),
|
||
|
MD5: md5,
|
||
|
CRC32C: crc32c,
|
||
|
MediaLink: o.MediaLink,
|
||
|
Metadata: o.Metadata,
|
||
|
Generation: o.Generation,
|
||
|
Metageneration: o.Metageneration,
|
||
|
StorageClass: o.StorageClass,
|
||
|
CustomerKeySHA256: sha256,
|
||
|
KMSKeyName: o.KmsKeyName,
|
||
|
Created: convertTime(o.TimeCreated),
|
||
|
Deleted: convertTime(o.TimeDeleted),
|
||
|
Updated: convertTime(o.Updated),
|
||
|
Etag: o.Etag,
|
||
|
CustomTime: convertTime(o.CustomTime),
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Decode a uint32 encoded in Base64 in big-endian byte order.
|
||
|
func decodeUint32(b64 string) (uint32, error) {
|
||
|
d, err := base64.StdEncoding.DecodeString(b64)
|
||
|
if err != nil {
|
||
|
return 0, err
|
||
|
}
|
||
|
if len(d) != 4 {
|
||
|
return 0, fmt.Errorf("storage: %q does not encode a 32-bit value", d)
|
||
|
}
|
||
|
return uint32(d[0])<<24 + uint32(d[1])<<16 + uint32(d[2])<<8 + uint32(d[3]), nil
|
||
|
}
|
||
|
|
||
|
// Encode a uint32 as Base64 in big-endian byte order.
|
||
|
func encodeUint32(u uint32) string {
|
||
|
b := []byte{byte(u >> 24), byte(u >> 16), byte(u >> 8), byte(u)}
|
||
|
return base64.StdEncoding.EncodeToString(b)
|
||
|
}
|
||
|
|
||
|
// Query represents a query to filter objects from a bucket.
|
||
|
type Query struct {
|
||
|
// Delimiter returns results in a directory-like fashion.
|
||
|
// Results will contain only objects whose names, aside from the
|
||
|
// prefix, do not contain delimiter. Objects whose names,
|
||
|
// aside from the prefix, contain delimiter will have their name,
|
||
|
// truncated after the delimiter, returned in prefixes.
|
||
|
// Duplicate prefixes are omitted.
|
||
|
// Optional.
|
||
|
Delimiter string
|
||
|
|
||
|
// Prefix is the prefix filter to query objects
|
||
|
// whose names begin with this prefix.
|
||
|
// Optional.
|
||
|
Prefix string
|
||
|
|
||
|
// Versions indicates whether multiple versions of the same
|
||
|
// object will be included in the results.
|
||
|
Versions bool
|
||
|
|
||
|
// fieldSelection is used to select only specific fields to be returned by
|
||
|
// the query. It's used internally and is populated for the user by
|
||
|
// calling Query.SetAttrSelection
|
||
|
fieldSelection string
|
||
|
|
||
|
// StartOffset is used to filter results to objects whose names are
|
||
|
// lexicographically equal to or after startOffset. If endOffset is also set,
|
||
|
// the objects listed will have names between startOffset (inclusive) and
|
||
|
// endOffset (exclusive).
|
||
|
StartOffset string
|
||
|
|
||
|
// EndOffset is used to filter results to objects whose names are
|
||
|
// lexicographically before endOffset. If startOffset is also set, the objects
|
||
|
// listed will have names between startOffset (inclusive) and endOffset (exclusive).
|
||
|
EndOffset string
|
||
|
}
|
||
|
|
||
|
// attrToFieldMap maps the field names of ObjectAttrs to the underlying field
|
||
|
// names in the API call. Only the ObjectAttrs field names are visible to users
|
||
|
// because they are already part of the public API of the package.
|
||
|
var attrToFieldMap = map[string]string{
|
||
|
"Bucket": "bucket",
|
||
|
"Name": "name",
|
||
|
"ContentType": "contentType",
|
||
|
"ContentLanguage": "contentLanguage",
|
||
|
"CacheControl": "cacheControl",
|
||
|
"EventBasedHold": "eventBasedHold",
|
||
|
"TemporaryHold": "temporaryHold",
|
||
|
"RetentionExpirationTime": "retentionExpirationTime",
|
||
|
"ACL": "acl",
|
||
|
"Owner": "owner",
|
||
|
"ContentEncoding": "contentEncoding",
|
||
|
"ContentDisposition": "contentDisposition",
|
||
|
"Size": "size",
|
||
|
"MD5": "md5Hash",
|
||
|
"CRC32C": "crc32c",
|
||
|
"MediaLink": "mediaLink",
|
||
|
"Metadata": "metadata",
|
||
|
"Generation": "generation",
|
||
|
"Metageneration": "metageneration",
|
||
|
"StorageClass": "storageClass",
|
||
|
"CustomerKeySHA256": "customerEncryption",
|
||
|
"KMSKeyName": "kmsKeyName",
|
||
|
"Created": "timeCreated",
|
||
|
"Deleted": "timeDeleted",
|
||
|
"Updated": "updated",
|
||
|
"Etag": "etag",
|
||
|
"CustomTime": "customTime",
|
||
|
}
|
||
|
|
||
|
// SetAttrSelection makes the query populate only specific attributes of
|
||
|
// objects. When iterating over objects, if you only need each object's name
|
||
|
// and size, pass []string{"Name", "Size"} to this method. Only these fields
|
||
|
// will be fetched for each object across the network; the other fields of
|
||
|
// ObjectAttr will remain at their default values. This is a performance
|
||
|
// optimization; for more information, see
|
||
|
// https://cloud.google.com/storage/docs/json_api/v1/how-tos/performance
|
||
|
func (q *Query) SetAttrSelection(attrs []string) error {
|
||
|
fieldSet := make(map[string]bool)
|
||
|
|
||
|
for _, attr := range attrs {
|
||
|
field, ok := attrToFieldMap[attr]
|
||
|
if !ok {
|
||
|
return fmt.Errorf("storage: attr %v is not valid", attr)
|
||
|
}
|
||
|
fieldSet[field] = true
|
||
|
}
|
||
|
|
||
|
if len(fieldSet) > 0 {
|
||
|
var b bytes.Buffer
|
||
|
b.WriteString("prefixes,items(")
|
||
|
first := true
|
||
|
for field := range fieldSet {
|
||
|
if !first {
|
||
|
b.WriteString(",")
|
||
|
}
|
||
|
first = false
|
||
|
b.WriteString(field)
|
||
|
}
|
||
|
b.WriteString(")")
|
||
|
q.fieldSelection = b.String()
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// Conditions constrain methods to act on specific generations of
|
||
|
// objects.
|
||
|
//
|
||
|
// The zero value is an empty set of constraints. Not all conditions or
|
||
|
// combinations of conditions are applicable to all methods.
|
||
|
// See https://cloud.google.com/storage/docs/generations-preconditions
|
||
|
// for details on how these operate.
|
||
|
type Conditions struct {
|
||
|
// Generation constraints.
|
||
|
// At most one of the following can be set to a non-zero value.
|
||
|
|
||
|
// GenerationMatch specifies that the object must have the given generation
|
||
|
// for the operation to occur.
|
||
|
// If GenerationMatch is zero, it has no effect.
|
||
|
// Use DoesNotExist to specify that the object does not exist in the bucket.
|
||
|
GenerationMatch int64
|
||
|
|
||
|
// GenerationNotMatch specifies that the object must not have the given
|
||
|
// generation for the operation to occur.
|
||
|
// If GenerationNotMatch is zero, it has no effect.
|
||
|
GenerationNotMatch int64
|
||
|
|
||
|
// DoesNotExist specifies that the object must not exist in the bucket for
|
||
|
// the operation to occur.
|
||
|
// If DoesNotExist is false, it has no effect.
|
||
|
DoesNotExist bool
|
||
|
|
||
|
// Metadata generation constraints.
|
||
|
// At most one of the following can be set to a non-zero value.
|
||
|
|
||
|
// MetagenerationMatch specifies that the object must have the given
|
||
|
// metageneration for the operation to occur.
|
||
|
// If MetagenerationMatch is zero, it has no effect.
|
||
|
MetagenerationMatch int64
|
||
|
|
||
|
// MetagenerationNotMatch specifies that the object must not have the given
|
||
|
// metageneration for the operation to occur.
|
||
|
// If MetagenerationNotMatch is zero, it has no effect.
|
||
|
MetagenerationNotMatch int64
|
||
|
}
|
||
|
|
||
|
func (c *Conditions) validate(method string) error {
|
||
|
if *c == (Conditions{}) {
|
||
|
return fmt.Errorf("storage: %s: empty conditions", method)
|
||
|
}
|
||
|
if !c.isGenerationValid() {
|
||
|
return fmt.Errorf("storage: %s: multiple conditions specified for generation", method)
|
||
|
}
|
||
|
if !c.isMetagenerationValid() {
|
||
|
return fmt.Errorf("storage: %s: multiple conditions specified for metageneration", method)
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func (c *Conditions) isGenerationValid() bool {
|
||
|
n := 0
|
||
|
if c.GenerationMatch != 0 {
|
||
|
n++
|
||
|
}
|
||
|
if c.GenerationNotMatch != 0 {
|
||
|
n++
|
||
|
}
|
||
|
if c.DoesNotExist {
|
||
|
n++
|
||
|
}
|
||
|
return n <= 1
|
||
|
}
|
||
|
|
||
|
func (c *Conditions) isMetagenerationValid() bool {
|
||
|
return c.MetagenerationMatch == 0 || c.MetagenerationNotMatch == 0
|
||
|
}
|
||
|
|
||
|
// applyConds modifies the provided call using the conditions in conds.
|
||
|
// call is something that quacks like a *raw.WhateverCall.
|
||
|
func applyConds(method string, gen int64, conds *Conditions, call interface{}) error {
|
||
|
cval := reflect.ValueOf(call)
|
||
|
if gen >= 0 {
|
||
|
if !setConditionField(cval, "Generation", gen) {
|
||
|
return fmt.Errorf("storage: %s: generation not supported", method)
|
||
|
}
|
||
|
}
|
||
|
if conds == nil {
|
||
|
return nil
|
||
|
}
|
||
|
if err := conds.validate(method); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
switch {
|
||
|
case conds.GenerationMatch != 0:
|
||
|
if !setConditionField(cval, "IfGenerationMatch", conds.GenerationMatch) {
|
||
|
return fmt.Errorf("storage: %s: ifGenerationMatch not supported", method)
|
||
|
}
|
||
|
case conds.GenerationNotMatch != 0:
|
||
|
if !setConditionField(cval, "IfGenerationNotMatch", conds.GenerationNotMatch) {
|
||
|
return fmt.Errorf("storage: %s: ifGenerationNotMatch not supported", method)
|
||
|
}
|
||
|
case conds.DoesNotExist:
|
||
|
if !setConditionField(cval, "IfGenerationMatch", int64(0)) {
|
||
|
return fmt.Errorf("storage: %s: DoesNotExist not supported", method)
|
||
|
}
|
||
|
}
|
||
|
switch {
|
||
|
case conds.MetagenerationMatch != 0:
|
||
|
if !setConditionField(cval, "IfMetagenerationMatch", conds.MetagenerationMatch) {
|
||
|
return fmt.Errorf("storage: %s: ifMetagenerationMatch not supported", method)
|
||
|
}
|
||
|
case conds.MetagenerationNotMatch != 0:
|
||
|
if !setConditionField(cval, "IfMetagenerationNotMatch", conds.MetagenerationNotMatch) {
|
||
|
return fmt.Errorf("storage: %s: ifMetagenerationNotMatch not supported", method)
|
||
|
}
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func applySourceConds(gen int64, conds *Conditions, call *raw.ObjectsRewriteCall) error {
|
||
|
if gen >= 0 {
|
||
|
call.SourceGeneration(gen)
|
||
|
}
|
||
|
if conds == nil {
|
||
|
return nil
|
||
|
}
|
||
|
if err := conds.validate("CopyTo source"); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
switch {
|
||
|
case conds.GenerationMatch != 0:
|
||
|
call.IfSourceGenerationMatch(conds.GenerationMatch)
|
||
|
case conds.GenerationNotMatch != 0:
|
||
|
call.IfSourceGenerationNotMatch(conds.GenerationNotMatch)
|
||
|
case conds.DoesNotExist:
|
||
|
call.IfSourceGenerationMatch(0)
|
||
|
}
|
||
|
switch {
|
||
|
case conds.MetagenerationMatch != 0:
|
||
|
call.IfSourceMetagenerationMatch(conds.MetagenerationMatch)
|
||
|
case conds.MetagenerationNotMatch != 0:
|
||
|
call.IfSourceMetagenerationNotMatch(conds.MetagenerationNotMatch)
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// setConditionField sets a field on a *raw.WhateverCall.
|
||
|
// We can't use anonymous interfaces because the return type is
|
||
|
// different, since the field setters are builders.
|
||
|
func setConditionField(call reflect.Value, name string, value interface{}) bool {
|
||
|
m := call.MethodByName(name)
|
||
|
if !m.IsValid() {
|
||
|
return false
|
||
|
}
|
||
|
m.Call([]reflect.Value{reflect.ValueOf(value)})
|
||
|
return true
|
||
|
}
|
||
|
|
||
|
// conditionsQuery returns the generation and conditions as a URL query
|
||
|
// string suitable for URL.RawQuery. It assumes that the conditions
|
||
|
// have been validated.
|
||
|
func conditionsQuery(gen int64, conds *Conditions) string {
|
||
|
// URL escapes are elided because integer strings are URL-safe.
|
||
|
var buf []byte
|
||
|
|
||
|
appendParam := func(s string, n int64) {
|
||
|
if len(buf) > 0 {
|
||
|
buf = append(buf, '&')
|
||
|
}
|
||
|
buf = append(buf, s...)
|
||
|
buf = strconv.AppendInt(buf, n, 10)
|
||
|
}
|
||
|
|
||
|
if gen >= 0 {
|
||
|
appendParam("generation=", gen)
|
||
|
}
|
||
|
if conds == nil {
|
||
|
return string(buf)
|
||
|
}
|
||
|
switch {
|
||
|
case conds.GenerationMatch != 0:
|
||
|
appendParam("ifGenerationMatch=", conds.GenerationMatch)
|
||
|
case conds.GenerationNotMatch != 0:
|
||
|
appendParam("ifGenerationNotMatch=", conds.GenerationNotMatch)
|
||
|
case conds.DoesNotExist:
|
||
|
appendParam("ifGenerationMatch=", 0)
|
||
|
}
|
||
|
switch {
|
||
|
case conds.MetagenerationMatch != 0:
|
||
|
appendParam("ifMetagenerationMatch=", conds.MetagenerationMatch)
|
||
|
case conds.MetagenerationNotMatch != 0:
|
||
|
appendParam("ifMetagenerationNotMatch=", conds.MetagenerationNotMatch)
|
||
|
}
|
||
|
return string(buf)
|
||
|
}
|
||
|
|
||
|
// composeSourceObj wraps a *raw.ComposeRequestSourceObjects, but adds the methods
|
||
|
// that modifyCall searches for by name.
|
||
|
type composeSourceObj struct {
|
||
|
src *raw.ComposeRequestSourceObjects
|
||
|
}
|
||
|
|
||
|
func (c composeSourceObj) Generation(gen int64) {
|
||
|
c.src.Generation = gen
|
||
|
}
|
||
|
|
||
|
func (c composeSourceObj) IfGenerationMatch(gen int64) {
|
||
|
// It's safe to overwrite ObjectPreconditions, since its only field is
|
||
|
// IfGenerationMatch.
|
||
|
c.src.ObjectPreconditions = &raw.ComposeRequestSourceObjectsObjectPreconditions{
|
||
|
IfGenerationMatch: gen,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func setEncryptionHeaders(headers http.Header, key []byte, copySource bool) error {
|
||
|
if key == nil {
|
||
|
return nil
|
||
|
}
|
||
|
// TODO(jbd): Ask the API team to return a more user-friendly error
|
||
|
// and avoid doing this check at the client level.
|
||
|
if len(key) != 32 {
|
||
|
return errors.New("storage: not a 32-byte AES-256 key")
|
||
|
}
|
||
|
var cs string
|
||
|
if copySource {
|
||
|
cs = "copy-source-"
|
||
|
}
|
||
|
headers.Set("x-goog-"+cs+"encryption-algorithm", "AES256")
|
||
|
headers.Set("x-goog-"+cs+"encryption-key", base64.StdEncoding.EncodeToString(key))
|
||
|
keyHash := sha256.Sum256(key)
|
||
|
headers.Set("x-goog-"+cs+"encryption-key-sha256", base64.StdEncoding.EncodeToString(keyHash[:]))
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// ServiceAccount fetches the email address of the given project's Google Cloud Storage service account.
|
||
|
func (c *Client) ServiceAccount(ctx context.Context, projectID string) (string, error) {
|
||
|
r := c.raw.Projects.ServiceAccount.Get(projectID)
|
||
|
res, err := r.Context(ctx).Do()
|
||
|
if err != nil {
|
||
|
return "", err
|
||
|
}
|
||
|
return res.EmailAddress, nil
|
||
|
}
|