# Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. require 'date' require 'google/apis/core/base_service' require 'google/apis/core/json_representation' require 'google/apis/core/hashable' require 'google/apis/errors' module Google module Apis module StoragetransferV1 # AWS access key (see [AWS Security Credentials](https://docs.aws.amazon.com/ # general/latest/gr/aws-security-credentials.html)). For information on our data # retention policy for user credentials, see [User credentials](data-retention# # user-credentials). class AwsAccessKey include Google::Apis::Core::Hashable # Required. AWS access key ID. # Corresponds to the JSON property `accessKeyId` # @return [String] attr_accessor :access_key_id # Required. AWS secret access key. This field is not returned in RPC responses. # Corresponds to the JSON property `secretAccessKey` # @return [String] attr_accessor :secret_access_key def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @access_key_id = args[:access_key_id] if args.key?(:access_key_id) @secret_access_key = args[:secret_access_key] if args.key?(:secret_access_key) end end # An AwsS3Data resource can be a data source, but not a data sink. In an # AwsS3Data resource, an object's name is the S3 object's key name. class AwsS3Data include Google::Apis::Core::Hashable # AWS access key (see [AWS Security Credentials](https://docs.aws.amazon.com/ # general/latest/gr/aws-security-credentials.html)). For information on our data # retention policy for user credentials, see [User credentials](data-retention# # user-credentials). # Corresponds to the JSON property `awsAccessKey` # @return [Google::Apis::StoragetransferV1::AwsAccessKey] attr_accessor :aws_access_key # Required. S3 Bucket name (see [Creating a bucket](https://docs.aws.amazon.com/ # AmazonS3/latest/dev/create-bucket-get-location-example.html)). # Corresponds to the JSON property `bucketName` # @return [String] attr_accessor :bucket_name # Root path to transfer objects. Must be an empty string or full path name that # ends with a '/'. This field is treated as an object prefix. As such, it should # generally not begin with a '/'. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @aws_access_key = args[:aws_access_key] if args.key?(:aws_access_key) @bucket_name = args[:bucket_name] if args.key?(:bucket_name) @path = args[:path] if args.key?(:path) end end # An AzureBlobStorageData resource can be a data source, but not a data sink. An # AzureBlobStorageData resource represents one Azure container. The storage # account determines the [Azure endpoint](https://docs.microsoft.com/en-us/azure/ # storage/common/storage-create-storage-account#storage-account-endpoints). In # an AzureBlobStorageData resource, a blobs's name is the [Azure Blob Storage # blob's key name](https://docs.microsoft.com/en-us/rest/api/storageservices/ # naming-and-referencing-containers--blobs--and-metadata#blob-names). class AzureBlobStorageData include Google::Apis::Core::Hashable # Azure credentials For information on our data retention policy for user # credentials, see [User credentials](data-retention#user-credentials). # Corresponds to the JSON property `azureCredentials` # @return [Google::Apis::StoragetransferV1::AzureCredentials] attr_accessor :azure_credentials # Required. The container to transfer from the Azure Storage account. # Corresponds to the JSON property `container` # @return [String] attr_accessor :container # Root path to transfer objects. Must be an empty string or full path name that # ends with a '/'. This field is treated as an object prefix. As such, it should # generally not begin with a '/'. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path # Required. The name of the Azure Storage account. # Corresponds to the JSON property `storageAccount` # @return [String] attr_accessor :storage_account def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @azure_credentials = args[:azure_credentials] if args.key?(:azure_credentials) @container = args[:container] if args.key?(:container) @path = args[:path] if args.key?(:path) @storage_account = args[:storage_account] if args.key?(:storage_account) end end # Azure credentials For information on our data retention policy for user # credentials, see [User credentials](data-retention#user-credentials). class AzureCredentials include Google::Apis::Core::Hashable # Required. Azure shared access signature. (see [Grant limited access to Azure # Storage resources using shared access signatures (SAS)](https://docs.microsoft. # com/en-us/azure/storage/common/storage-sas-overview)). # Corresponds to the JSON property `sasToken` # @return [String] attr_accessor :sas_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @sas_token = args[:sas_token] if args.key?(:sas_token) end end # The request message for Operations.CancelOperation. class CancelOperationRequest include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # Represents a whole or partial calendar date, such as a birthday. The time of # day and time zone are either specified elsewhere or are insignificant. The # date is relative to the Gregorian Calendar. This can represent one of the # following: * A full date, with non-zero year, month, and day values * A month # and day value, with a zero year, such as an anniversary * A year on its own, # with zero month and day values * A year and month value, with a zero day, such # as a credit card expiration date Related types are google.type.TimeOfDay and ` # google.protobuf.Timestamp`. class Date include Google::Apis::Core::Hashable # Day of a month. Must be from 1 to 31 and valid for the year and month, or 0 to # specify a year by itself or a year and month where the day isn't significant. # Corresponds to the JSON property `day` # @return [Fixnum] attr_accessor :day # Month of a year. Must be from 1 to 12, or 0 to specify a year without a month # and day. # Corresponds to the JSON property `month` # @return [Fixnum] attr_accessor :month # Year of the date. Must be from 1 to 9999, or 0 to specify a date without a # year. # Corresponds to the JSON property `year` # @return [Fixnum] attr_accessor :year def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @day = args[:day] if args.key?(:day) @month = args[:month] if args.key?(:month) @year = args[:year] if args.key?(:year) end end # A generic empty message that you can re-use to avoid defining duplicated empty # messages in your APIs. A typical example is to use it as the request or the # response type of an API method. For instance: service Foo ` rpc Bar(google. # protobuf.Empty) returns (google.protobuf.Empty); ` The JSON representation for # `Empty` is empty JSON object ````. class Empty include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # An entry describing an error that has occurred. class ErrorLogEntry include Google::Apis::Core::Hashable # A list of messages that carry the error details. # Corresponds to the JSON property `errorDetails` # @return [Array] attr_accessor :error_details # Required. A URL that refers to the target (a data source, a data sink, or an # object) with which the error is associated. # Corresponds to the JSON property `url` # @return [String] attr_accessor :url def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @error_details = args[:error_details] if args.key?(:error_details) @url = args[:url] if args.key?(:url) end end # A summary of errors by error code, plus a count and sample error log entries. class ErrorSummary include Google::Apis::Core::Hashable # Required. # Corresponds to the JSON property `errorCode` # @return [String] attr_accessor :error_code # Required. Count of this type of error. # Corresponds to the JSON property `errorCount` # @return [Fixnum] attr_accessor :error_count # Error samples. At most 5 error log entries will be recorded for a given error # code for a single transfer operation. # Corresponds to the JSON property `errorLogEntries` # @return [Array] attr_accessor :error_log_entries def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @error_code = args[:error_code] if args.key?(:error_code) @error_count = args[:error_count] if args.key?(:error_count) @error_log_entries = args[:error_log_entries] if args.key?(:error_log_entries) end end # In a GcsData resource, an object's name is the Cloud Storage object's name and # its "last modification time" refers to the object's `updated` property of # Cloud Storage objects, which changes when the content or the metadata of the # object is updated. class GcsData include Google::Apis::Core::Hashable # Required. Cloud Storage bucket name (see [Bucket Name Requirements](https:// # cloud.google.com/storage/docs/naming#requirements)). # Corresponds to the JSON property `bucketName` # @return [String] attr_accessor :bucket_name # Root path to transfer objects. Must be an empty string or full path name that # ends with a '/'. This field is treated as an object prefix. As such, it should # generally not begin with a '/'. (must meet Object Name Requirements](https:// # cloud.google.com/storage/docs/naming#objectnames)). # Corresponds to the JSON property `path` # @return [String] attr_accessor :path def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @bucket_name = args[:bucket_name] if args.key?(:bucket_name) @path = args[:path] if args.key?(:path) end end # Google service account class GoogleServiceAccount include Google::Apis::Core::Hashable # Email address of the service account. # Corresponds to the JSON property `accountEmail` # @return [String] attr_accessor :account_email def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @account_email = args[:account_email] if args.key?(:account_email) end end # An HttpData resource specifies a list of objects on the web to be transferred # over HTTP. The information of the objects to be transferred is contained in a # file referenced by a URL. The first line in the file must be `"TsvHttpData-1.0" # `, which specifies the format of the file. Subsequent lines specify the # information of the list of objects, one object per list entry. Each entry has # the following tab-delimited fields: * **HTTP URL** — The location of the # object. * **Length** — The size of the object in bytes. * **MD5** — The base64- # encoded MD5 hash of the object. For an example of a valid TSV file, see [ # Transferring data from URLs](https://cloud.google.com/storage-transfer/docs/ # create-url-list). When transferring data based on a URL list, keep the # following in mind: * When an object located at `http(s)://hostname:port/` is # transferred to a data sink, the name of the object at the data sink is `/`. * # If the specified size of an object does not match the actual size of the # object fetched, the object will not be transferred. * If the specified MD5 # does not match the MD5 computed from the transferred bytes, the object # transfer will fail. * Ensure that each URL you specify is publicly accessible. # For example, in Cloud Storage you can [share an object publicly] (https:// # cloud.google.com/storage/docs/cloud-console#_sharingdata) and get a link to it. # * Storage Transfer Service obeys `robots.txt` rules and requires the source # HTTP server to support `Range` requests and to return a `Content-Length` # header in each response. * ObjectConditions have no effect when filtering # objects to transfer. class HttpData include Google::Apis::Core::Hashable # Required. The URL that points to the file that stores the object list entries. # This file must allow public access. Currently, only URLs with HTTP and HTTPS # schemes are supported. # Corresponds to the JSON property `listUrl` # @return [String] attr_accessor :list_url def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @list_url = args[:list_url] if args.key?(:list_url) end end # The response message for Operations.ListOperations. class ListOperationsResponse include Google::Apis::Core::Hashable # The standard List next-page token. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # A list of operations that matches the specified filter in the request. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # Response from ListTransferJobs. class ListTransferJobsResponse include Google::Apis::Core::Hashable # The list next page token. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # A list of transfer jobs. # Corresponds to the JSON property `transferJobs` # @return [Array] attr_accessor :transfer_jobs def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @transfer_jobs = args[:transfer_jobs] if args.key?(:transfer_jobs) end end # Specification to configure notifications published to Cloud Pub/Sub. # Notifications will be published to the customer-provided topic using the # following `PubsubMessage.attributes`: * `"eventType"`: one of the EventType # values * `"payloadFormat"`: one of the PayloadFormat values * `"projectId"`: # the project_id of the `TransferOperation` * `"transferJobName"`: the # transfer_job_name of the `TransferOperation` * `"transferOperationName"`: the # name of the `TransferOperation` The `PubsubMessage.data` will contain a # TransferOperation resource formatted according to the specified `PayloadFormat` # . class NotificationConfig include Google::Apis::Core::Hashable # Event types for which a notification is desired. If empty, send notifications # for all event types. # Corresponds to the JSON property `eventTypes` # @return [Array] attr_accessor :event_types # Required. The desired format of the notification message payloads. # Corresponds to the JSON property `payloadFormat` # @return [String] attr_accessor :payload_format # Required. The `Topic.name` of the Cloud Pub/Sub topic to which to publish # notifications. Must be of the format: `projects/`project`/topics/`topic``. Not # matching this format will result in an INVALID_ARGUMENT error. # Corresponds to the JSON property `pubsubTopic` # @return [String] attr_accessor :pubsub_topic def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @event_types = args[:event_types] if args.key?(:event_types) @payload_format = args[:payload_format] if args.key?(:payload_format) @pubsub_topic = args[:pubsub_topic] if args.key?(:pubsub_topic) end end # Conditions that determine which objects will be transferred. Applies only to # Cloud Data Sources such as S3, Azure, and Cloud Storage. The "last # modification time" refers to the time of the last change to the object's # content or metadata — specifically, this is the `updated` property of Cloud # Storage objects, the `LastModified` field of S3 objects, and the `Last- # Modified` header of Azure blobs. class ObjectConditions include Google::Apis::Core::Hashable # `exclude_prefixes` must follow the requirements described for include_prefixes. # The max size of `exclude_prefixes` is 1000. # Corresponds to the JSON property `excludePrefixes` # @return [Array] attr_accessor :exclude_prefixes # If `include_prefixes` is specified, objects that satisfy the object conditions # must have names that start with one of the `include_prefixes` and that do not # start with any of the exclude_prefixes. If `include_prefixes` is not specified, # all objects except those that have names starting with one of the ` # exclude_prefixes` must satisfy the object conditions. Requirements: * Each # include-prefix and exclude-prefix can contain any sequence of Unicode # characters, to a max length of 1024 bytes when UTF8-encoded, and must not # contain Carriage Return or Line Feed characters. Wildcard matching and regular # expression matching are not supported. * Each include-prefix and exclude- # prefix must omit the leading slash. For example, to include the `requests.gz` # object in a transfer from `s3://my-aws-bucket/logs/y=2015/requests.gz`, # specify the include prefix as `logs/y=2015/requests.gz`. * None of the include- # prefix or the exclude-prefix values can be empty, if specified. * Each include- # prefix must include a distinct portion of the object namespace. No include- # prefix may be a prefix of another include-prefix. * Each exclude-prefix must # exclude a distinct portion of the object namespace. No exclude-prefix may be a # prefix of another exclude-prefix. * If `include_prefixes` is specified, then # each exclude-prefix must start with the value of a path explicitly included by # `include_prefixes`. The max size of `include_prefixes` is 1000. # Corresponds to the JSON property `includePrefixes` # @return [Array] attr_accessor :include_prefixes # If specified, only objects with a "last modification time" before this # timestamp and objects that don't have a "last modification time" will be # transferred. # Corresponds to the JSON property `lastModifiedBefore` # @return [String] attr_accessor :last_modified_before # If specified, only objects with a "last modification time" on or after this # timestamp and objects that don't have a "last modification time" are # transferred. The `last_modified_since` and `last_modified_before` fields can # be used together for chunked data processing. For example, consider a script # that processes each day's worth of data at a time. For that you'd set each of # the fields as follows: * `last_modified_since` to the start of the day * ` # last_modified_before` to the end of the day # Corresponds to the JSON property `lastModifiedSince` # @return [String] attr_accessor :last_modified_since # If specified, only objects with a "last modification time" on or after `NOW` - # `max_time_elapsed_since_last_modification` and objects that don't have a "last # modification time" are transferred. For each TransferOperation started by this # TransferJob, `NOW` refers to the start_time of the `TransferOperation`. # Corresponds to the JSON property `maxTimeElapsedSinceLastModification` # @return [String] attr_accessor :max_time_elapsed_since_last_modification # If specified, only objects with a "last modification time" before `NOW` - ` # min_time_elapsed_since_last_modification` and objects that don't have a "last # modification time" are transferred. For each TransferOperation started by this # TransferJob, `NOW` refers to the start_time of the `TransferOperation`. # Corresponds to the JSON property `minTimeElapsedSinceLastModification` # @return [String] attr_accessor :min_time_elapsed_since_last_modification def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @exclude_prefixes = args[:exclude_prefixes] if args.key?(:exclude_prefixes) @include_prefixes = args[:include_prefixes] if args.key?(:include_prefixes) @last_modified_before = args[:last_modified_before] if args.key?(:last_modified_before) @last_modified_since = args[:last_modified_since] if args.key?(:last_modified_since) @max_time_elapsed_since_last_modification = args[:max_time_elapsed_since_last_modification] if args.key?(:max_time_elapsed_since_last_modification) @min_time_elapsed_since_last_modification = args[:min_time_elapsed_since_last_modification] if args.key?(:min_time_elapsed_since_last_modification) end end # This resource represents a long-running operation that is the result of a # network API call. class Operation include Google::Apis::Core::Hashable # If the value is `false`, it means the operation is still in progress. If `true` # , the operation is completed, and either `error` or `response` is available. # Corresponds to the JSON property `done` # @return [Boolean] attr_accessor :done alias_method :done?, :done # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `error` # @return [Google::Apis::StoragetransferV1::Status] attr_accessor :error # Represents the transfer operation object. To request a TransferOperation # object, use transferOperations.get. # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # The server-assigned name, which is only unique within the same service that # originally returns it. If you use the default HTTP mapping, the `name` should # have the format of `transferOperations/some/unique/name`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The normal response of the operation in case of success. If the original # method returns no data on success, such as `Delete`, the response is `google. # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, # the response should be the resource. For other methods, the response should # have the type `XxxResponse`, where `Xxx` is the original method name. For # example, if the original method name is `TakeSnapshot()`, the inferred # response type is `TakeSnapshotResponse`. # Corresponds to the JSON property `response` # @return [Hash] attr_accessor :response def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @done = args[:done] if args.key?(:done) @error = args[:error] if args.key?(:error) @metadata = args[:metadata] if args.key?(:metadata) @name = args[:name] if args.key?(:name) @response = args[:response] if args.key?(:response) end end # Request passed to PauseTransferOperation. class PauseTransferOperationRequest include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # Request passed to ResumeTransferOperation. class ResumeTransferOperationRequest include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # Transfers can be scheduled to recur or to run just once. class Schedule include Google::Apis::Core::Hashable # Represents a time of day. The date and time zone are either not significant or # are specified elsewhere. An API may choose to allow leap seconds. Related # types are google.type.Date and `google.protobuf.Timestamp`. # Corresponds to the JSON property `endTimeOfDay` # @return [Google::Apis::StoragetransferV1::TimeOfDay] attr_accessor :end_time_of_day # Interval between the start of each scheduled TransferOperation. If unspecified, # the default value is 24 hours. This value may not be less than 1 hour. # Corresponds to the JSON property `repeatInterval` # @return [String] attr_accessor :repeat_interval # Represents a whole or partial calendar date, such as a birthday. The time of # day and time zone are either specified elsewhere or are insignificant. The # date is relative to the Gregorian Calendar. This can represent one of the # following: * A full date, with non-zero year, month, and day values * A month # and day value, with a zero year, such as an anniversary * A year on its own, # with zero month and day values * A year and month value, with a zero day, such # as a credit card expiration date Related types are google.type.TimeOfDay and ` # google.protobuf.Timestamp`. # Corresponds to the JSON property `scheduleEndDate` # @return [Google::Apis::StoragetransferV1::Date] attr_accessor :schedule_end_date # Represents a whole or partial calendar date, such as a birthday. The time of # day and time zone are either specified elsewhere or are insignificant. The # date is relative to the Gregorian Calendar. This can represent one of the # following: * A full date, with non-zero year, month, and day values * A month # and day value, with a zero year, such as an anniversary * A year on its own, # with zero month and day values * A year and month value, with a zero day, such # as a credit card expiration date Related types are google.type.TimeOfDay and ` # google.protobuf.Timestamp`. # Corresponds to the JSON property `scheduleStartDate` # @return [Google::Apis::StoragetransferV1::Date] attr_accessor :schedule_start_date # Represents a time of day. The date and time zone are either not significant or # are specified elsewhere. An API may choose to allow leap seconds. Related # types are google.type.Date and `google.protobuf.Timestamp`. # Corresponds to the JSON property `startTimeOfDay` # @return [Google::Apis::StoragetransferV1::TimeOfDay] attr_accessor :start_time_of_day def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @end_time_of_day = args[:end_time_of_day] if args.key?(:end_time_of_day) @repeat_interval = args[:repeat_interval] if args.key?(:repeat_interval) @schedule_end_date = args[:schedule_end_date] if args.key?(:schedule_end_date) @schedule_start_date = args[:schedule_start_date] if args.key?(:schedule_start_date) @start_time_of_day = args[:start_time_of_day] if args.key?(:start_time_of_day) end end # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). class Status include Google::Apis::Core::Hashable # The status code, which should be an enum value of google.rpc.Code. # Corresponds to the JSON property `code` # @return [Fixnum] attr_accessor :code # A list of messages that carry the error details. There is a common set of # message types for APIs to use. # Corresponds to the JSON property `details` # @return [Array>] attr_accessor :details # A developer-facing error message, which should be in English. Any user-facing # error message should be localized and sent in the google.rpc.Status.details # field, or localized by the client. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @code = args[:code] if args.key?(:code) @details = args[:details] if args.key?(:details) @message = args[:message] if args.key?(:message) end end # Represents a time of day. The date and time zone are either not significant or # are specified elsewhere. An API may choose to allow leap seconds. Related # types are google.type.Date and `google.protobuf.Timestamp`. class TimeOfDay include Google::Apis::Core::Hashable # Hours of day in 24 hour format. Should be from 0 to 23. An API may choose to # allow the value "24:00:00" for scenarios like business closing time. # Corresponds to the JSON property `hours` # @return [Fixnum] attr_accessor :hours # Minutes of hour of day. Must be from 0 to 59. # Corresponds to the JSON property `minutes` # @return [Fixnum] attr_accessor :minutes # Fractions of seconds in nanoseconds. Must be from 0 to 999,999,999. # Corresponds to the JSON property `nanos` # @return [Fixnum] attr_accessor :nanos # Seconds of minutes of the time. Must normally be from 0 to 59. An API may # allow the value 60 if it allows leap-seconds. # Corresponds to the JSON property `seconds` # @return [Fixnum] attr_accessor :seconds def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @hours = args[:hours] if args.key?(:hours) @minutes = args[:minutes] if args.key?(:minutes) @nanos = args[:nanos] if args.key?(:nanos) @seconds = args[:seconds] if args.key?(:seconds) end end # A collection of counters that report the progress of a transfer operation. class TransferCounters include Google::Apis::Core::Hashable # Bytes that are copied to the data sink. # Corresponds to the JSON property `bytesCopiedToSink` # @return [Fixnum] attr_accessor :bytes_copied_to_sink # Bytes that are deleted from the data sink. # Corresponds to the JSON property `bytesDeletedFromSink` # @return [Fixnum] attr_accessor :bytes_deleted_from_sink # Bytes that are deleted from the data source. # Corresponds to the JSON property `bytesDeletedFromSource` # @return [Fixnum] attr_accessor :bytes_deleted_from_source # Bytes that failed to be deleted from the data sink. # Corresponds to the JSON property `bytesFailedToDeleteFromSink` # @return [Fixnum] attr_accessor :bytes_failed_to_delete_from_sink # Bytes found in the data source that are scheduled to be transferred, excluding # any that are filtered based on object conditions or skipped due to sync. # Corresponds to the JSON property `bytesFoundFromSource` # @return [Fixnum] attr_accessor :bytes_found_from_source # Bytes found only in the data sink that are scheduled to be deleted. # Corresponds to the JSON property `bytesFoundOnlyFromSink` # @return [Fixnum] attr_accessor :bytes_found_only_from_sink # Bytes in the data source that failed to be transferred or that failed to be # deleted after being transferred. # Corresponds to the JSON property `bytesFromSourceFailed` # @return [Fixnum] attr_accessor :bytes_from_source_failed # Bytes in the data source that are not transferred because they already exist # in the data sink. # Corresponds to the JSON property `bytesFromSourceSkippedBySync` # @return [Fixnum] attr_accessor :bytes_from_source_skipped_by_sync # Objects that are copied to the data sink. # Corresponds to the JSON property `objectsCopiedToSink` # @return [Fixnum] attr_accessor :objects_copied_to_sink # Objects that are deleted from the data sink. # Corresponds to the JSON property `objectsDeletedFromSink` # @return [Fixnum] attr_accessor :objects_deleted_from_sink # Objects that are deleted from the data source. # Corresponds to the JSON property `objectsDeletedFromSource` # @return [Fixnum] attr_accessor :objects_deleted_from_source # Objects that failed to be deleted from the data sink. # Corresponds to the JSON property `objectsFailedToDeleteFromSink` # @return [Fixnum] attr_accessor :objects_failed_to_delete_from_sink # Objects found in the data source that are scheduled to be transferred, # excluding any that are filtered based on object conditions or skipped due to # sync. # Corresponds to the JSON property `objectsFoundFromSource` # @return [Fixnum] attr_accessor :objects_found_from_source # Objects found only in the data sink that are scheduled to be deleted. # Corresponds to the JSON property `objectsFoundOnlyFromSink` # @return [Fixnum] attr_accessor :objects_found_only_from_sink # Objects in the data source that failed to be transferred or that failed to be # deleted after being transferred. # Corresponds to the JSON property `objectsFromSourceFailed` # @return [Fixnum] attr_accessor :objects_from_source_failed # Objects in the data source that are not transferred because they already exist # in the data sink. # Corresponds to the JSON property `objectsFromSourceSkippedBySync` # @return [Fixnum] attr_accessor :objects_from_source_skipped_by_sync def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @bytes_copied_to_sink = args[:bytes_copied_to_sink] if args.key?(:bytes_copied_to_sink) @bytes_deleted_from_sink = args[:bytes_deleted_from_sink] if args.key?(:bytes_deleted_from_sink) @bytes_deleted_from_source = args[:bytes_deleted_from_source] if args.key?(:bytes_deleted_from_source) @bytes_failed_to_delete_from_sink = args[:bytes_failed_to_delete_from_sink] if args.key?(:bytes_failed_to_delete_from_sink) @bytes_found_from_source = args[:bytes_found_from_source] if args.key?(:bytes_found_from_source) @bytes_found_only_from_sink = args[:bytes_found_only_from_sink] if args.key?(:bytes_found_only_from_sink) @bytes_from_source_failed = args[:bytes_from_source_failed] if args.key?(:bytes_from_source_failed) @bytes_from_source_skipped_by_sync = args[:bytes_from_source_skipped_by_sync] if args.key?(:bytes_from_source_skipped_by_sync) @objects_copied_to_sink = args[:objects_copied_to_sink] if args.key?(:objects_copied_to_sink) @objects_deleted_from_sink = args[:objects_deleted_from_sink] if args.key?(:objects_deleted_from_sink) @objects_deleted_from_source = args[:objects_deleted_from_source] if args.key?(:objects_deleted_from_source) @objects_failed_to_delete_from_sink = args[:objects_failed_to_delete_from_sink] if args.key?(:objects_failed_to_delete_from_sink) @objects_found_from_source = args[:objects_found_from_source] if args.key?(:objects_found_from_source) @objects_found_only_from_sink = args[:objects_found_only_from_sink] if args.key?(:objects_found_only_from_sink) @objects_from_source_failed = args[:objects_from_source_failed] if args.key?(:objects_from_source_failed) @objects_from_source_skipped_by_sync = args[:objects_from_source_skipped_by_sync] if args.key?(:objects_from_source_skipped_by_sync) end end # This resource represents the configuration of a transfer job that runs # periodically. class TransferJob include Google::Apis::Core::Hashable # Output only. The time that the transfer job was created. # Corresponds to the JSON property `creationTime` # @return [String] attr_accessor :creation_time # Output only. The time that the transfer job was deleted. # Corresponds to the JSON property `deletionTime` # @return [String] attr_accessor :deletion_time # A description provided by the user for the job. Its max length is 1024 bytes # when Unicode-encoded. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description # Output only. The time that the transfer job was last modified. # Corresponds to the JSON property `lastModificationTime` # @return [String] attr_accessor :last_modification_time # The name of the most recently started TransferOperation of this JobConfig. # Present if and only if at least one TransferOperation has been created for # this JobConfig. # Corresponds to the JSON property `latestOperationName` # @return [String] attr_accessor :latest_operation_name # A unique name (within the transfer project) assigned when the job is created. # If this field is empty in a CreateTransferJobRequest, Storage Transfer Service # will assign a unique name. Otherwise, the specified name is used as the unique # name for this job. If the specified name is in use by a job, the creation # request fails with an ALREADY_EXISTS error. This name must start with `" # transferJobs/"` prefix and end with a letter or a number, and should be no # more than 128 characters. Example: `"transferJobs/[A-Za-z0-9-._~]*[A-Za-z0-9]$" # ` Invalid job names will fail with an INVALID_ARGUMENT error. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Specification to configure notifications published to Cloud Pub/Sub. # Notifications will be published to the customer-provided topic using the # following `PubsubMessage.attributes`: * `"eventType"`: one of the EventType # values * `"payloadFormat"`: one of the PayloadFormat values * `"projectId"`: # the project_id of the `TransferOperation` * `"transferJobName"`: the # transfer_job_name of the `TransferOperation` * `"transferOperationName"`: the # name of the `TransferOperation` The `PubsubMessage.data` will contain a # TransferOperation resource formatted according to the specified `PayloadFormat` # . # Corresponds to the JSON property `notificationConfig` # @return [Google::Apis::StoragetransferV1::NotificationConfig] attr_accessor :notification_config # The ID of the Google Cloud Platform Project that owns the job. # Corresponds to the JSON property `projectId` # @return [String] attr_accessor :project_id # Transfers can be scheduled to recur or to run just once. # Corresponds to the JSON property `schedule` # @return [Google::Apis::StoragetransferV1::Schedule] attr_accessor :schedule # Status of the job. This value MUST be specified for `CreateTransferJobRequests` # . **Note:** The effect of the new job status takes place during a subsequent # job run. For example, if you change the job status from ENABLED to DISABLED, # and an operation spawned by the transfer is running, the status change would # not affect the current operation. # Corresponds to the JSON property `status` # @return [String] attr_accessor :status # Configuration for running a transfer. # Corresponds to the JSON property `transferSpec` # @return [Google::Apis::StoragetransferV1::TransferSpec] attr_accessor :transfer_spec def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @creation_time = args[:creation_time] if args.key?(:creation_time) @deletion_time = args[:deletion_time] if args.key?(:deletion_time) @description = args[:description] if args.key?(:description) @last_modification_time = args[:last_modification_time] if args.key?(:last_modification_time) @latest_operation_name = args[:latest_operation_name] if args.key?(:latest_operation_name) @name = args[:name] if args.key?(:name) @notification_config = args[:notification_config] if args.key?(:notification_config) @project_id = args[:project_id] if args.key?(:project_id) @schedule = args[:schedule] if args.key?(:schedule) @status = args[:status] if args.key?(:status) @transfer_spec = args[:transfer_spec] if args.key?(:transfer_spec) end end # A description of the execution of a transfer. class TransferOperation include Google::Apis::Core::Hashable # A collection of counters that report the progress of a transfer operation. # Corresponds to the JSON property `counters` # @return [Google::Apis::StoragetransferV1::TransferCounters] attr_accessor :counters # End time of this transfer execution. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # Summarizes errors encountered with sample error log entries. # Corresponds to the JSON property `errorBreakdowns` # @return [Array] attr_accessor :error_breakdowns # A globally unique ID assigned by the system. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Specification to configure notifications published to Cloud Pub/Sub. # Notifications will be published to the customer-provided topic using the # following `PubsubMessage.attributes`: * `"eventType"`: one of the EventType # values * `"payloadFormat"`: one of the PayloadFormat values * `"projectId"`: # the project_id of the `TransferOperation` * `"transferJobName"`: the # transfer_job_name of the `TransferOperation` * `"transferOperationName"`: the # name of the `TransferOperation` The `PubsubMessage.data` will contain a # TransferOperation resource formatted according to the specified `PayloadFormat` # . # Corresponds to the JSON property `notificationConfig` # @return [Google::Apis::StoragetransferV1::NotificationConfig] attr_accessor :notification_config # The ID of the Google Cloud Platform Project that owns the operation. # Corresponds to the JSON property `projectId` # @return [String] attr_accessor :project_id # Start time of this transfer execution. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time # Status of the transfer operation. # Corresponds to the JSON property `status` # @return [String] attr_accessor :status # The name of the transfer job that triggers this transfer operation. # Corresponds to the JSON property `transferJobName` # @return [String] attr_accessor :transfer_job_name # Configuration for running a transfer. # Corresponds to the JSON property `transferSpec` # @return [Google::Apis::StoragetransferV1::TransferSpec] attr_accessor :transfer_spec def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @counters = args[:counters] if args.key?(:counters) @end_time = args[:end_time] if args.key?(:end_time) @error_breakdowns = args[:error_breakdowns] if args.key?(:error_breakdowns) @name = args[:name] if args.key?(:name) @notification_config = args[:notification_config] if args.key?(:notification_config) @project_id = args[:project_id] if args.key?(:project_id) @start_time = args[:start_time] if args.key?(:start_time) @status = args[:status] if args.key?(:status) @transfer_job_name = args[:transfer_job_name] if args.key?(:transfer_job_name) @transfer_spec = args[:transfer_spec] if args.key?(:transfer_spec) end end # TransferOptions define the actions to be performed on objects in a transfer. class TransferOptions include Google::Apis::Core::Hashable # Whether objects should be deleted from the source after they are transferred # to the sink. **Note:** This option and delete_objects_unique_in_sink are # mutually exclusive. # Corresponds to the JSON property `deleteObjectsFromSourceAfterTransfer` # @return [Boolean] attr_accessor :delete_objects_from_source_after_transfer alias_method :delete_objects_from_source_after_transfer?, :delete_objects_from_source_after_transfer # Whether objects that exist only in the sink should be deleted. **Note:** This # option and delete_objects_from_source_after_transfer are mutually exclusive. # Corresponds to the JSON property `deleteObjectsUniqueInSink` # @return [Boolean] attr_accessor :delete_objects_unique_in_sink alias_method :delete_objects_unique_in_sink?, :delete_objects_unique_in_sink # When to overwrite objects that already exist in the sink. The default is that # only objects that are different from the source are ovewritten. If true, all # objects in the sink whose name matches an object in the source will be # overwritten with the source object. # Corresponds to the JSON property `overwriteObjectsAlreadyExistingInSink` # @return [Boolean] attr_accessor :overwrite_objects_already_existing_in_sink alias_method :overwrite_objects_already_existing_in_sink?, :overwrite_objects_already_existing_in_sink def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @delete_objects_from_source_after_transfer = args[:delete_objects_from_source_after_transfer] if args.key?(:delete_objects_from_source_after_transfer) @delete_objects_unique_in_sink = args[:delete_objects_unique_in_sink] if args.key?(:delete_objects_unique_in_sink) @overwrite_objects_already_existing_in_sink = args[:overwrite_objects_already_existing_in_sink] if args.key?(:overwrite_objects_already_existing_in_sink) end end # Configuration for running a transfer. class TransferSpec include Google::Apis::Core::Hashable # An AwsS3Data resource can be a data source, but not a data sink. In an # AwsS3Data resource, an object's name is the S3 object's key name. # Corresponds to the JSON property `awsS3DataSource` # @return [Google::Apis::StoragetransferV1::AwsS3Data] attr_accessor :aws_s3_data_source # An AzureBlobStorageData resource can be a data source, but not a data sink. An # AzureBlobStorageData resource represents one Azure container. The storage # account determines the [Azure endpoint](https://docs.microsoft.com/en-us/azure/ # storage/common/storage-create-storage-account#storage-account-endpoints). In # an AzureBlobStorageData resource, a blobs's name is the [Azure Blob Storage # blob's key name](https://docs.microsoft.com/en-us/rest/api/storageservices/ # naming-and-referencing-containers--blobs--and-metadata#blob-names). # Corresponds to the JSON property `azureBlobStorageDataSource` # @return [Google::Apis::StoragetransferV1::AzureBlobStorageData] attr_accessor :azure_blob_storage_data_source # In a GcsData resource, an object's name is the Cloud Storage object's name and # its "last modification time" refers to the object's `updated` property of # Cloud Storage objects, which changes when the content or the metadata of the # object is updated. # Corresponds to the JSON property `gcsDataSink` # @return [Google::Apis::StoragetransferV1::GcsData] attr_accessor :gcs_data_sink # In a GcsData resource, an object's name is the Cloud Storage object's name and # its "last modification time" refers to the object's `updated` property of # Cloud Storage objects, which changes when the content or the metadata of the # object is updated. # Corresponds to the JSON property `gcsDataSource` # @return [Google::Apis::StoragetransferV1::GcsData] attr_accessor :gcs_data_source # An HttpData resource specifies a list of objects on the web to be transferred # over HTTP. The information of the objects to be transferred is contained in a # file referenced by a URL. The first line in the file must be `"TsvHttpData-1.0" # `, which specifies the format of the file. Subsequent lines specify the # information of the list of objects, one object per list entry. Each entry has # the following tab-delimited fields: * **HTTP URL** — The location of the # object. * **Length** — The size of the object in bytes. * **MD5** — The base64- # encoded MD5 hash of the object. For an example of a valid TSV file, see [ # Transferring data from URLs](https://cloud.google.com/storage-transfer/docs/ # create-url-list). When transferring data based on a URL list, keep the # following in mind: * When an object located at `http(s)://hostname:port/` is # transferred to a data sink, the name of the object at the data sink is `/`. * # If the specified size of an object does not match the actual size of the # object fetched, the object will not be transferred. * If the specified MD5 # does not match the MD5 computed from the transferred bytes, the object # transfer will fail. * Ensure that each URL you specify is publicly accessible. # For example, in Cloud Storage you can [share an object publicly] (https:// # cloud.google.com/storage/docs/cloud-console#_sharingdata) and get a link to it. # * Storage Transfer Service obeys `robots.txt` rules and requires the source # HTTP server to support `Range` requests and to return a `Content-Length` # header in each response. * ObjectConditions have no effect when filtering # objects to transfer. # Corresponds to the JSON property `httpDataSource` # @return [Google::Apis::StoragetransferV1::HttpData] attr_accessor :http_data_source # Conditions that determine which objects will be transferred. Applies only to # Cloud Data Sources such as S3, Azure, and Cloud Storage. The "last # modification time" refers to the time of the last change to the object's # content or metadata — specifically, this is the `updated` property of Cloud # Storage objects, the `LastModified` field of S3 objects, and the `Last- # Modified` header of Azure blobs. # Corresponds to the JSON property `objectConditions` # @return [Google::Apis::StoragetransferV1::ObjectConditions] attr_accessor :object_conditions # TransferOptions define the actions to be performed on objects in a transfer. # Corresponds to the JSON property `transferOptions` # @return [Google::Apis::StoragetransferV1::TransferOptions] attr_accessor :transfer_options def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @aws_s3_data_source = args[:aws_s3_data_source] if args.key?(:aws_s3_data_source) @azure_blob_storage_data_source = args[:azure_blob_storage_data_source] if args.key?(:azure_blob_storage_data_source) @gcs_data_sink = args[:gcs_data_sink] if args.key?(:gcs_data_sink) @gcs_data_source = args[:gcs_data_source] if args.key?(:gcs_data_source) @http_data_source = args[:http_data_source] if args.key?(:http_data_source) @object_conditions = args[:object_conditions] if args.key?(:object_conditions) @transfer_options = args[:transfer_options] if args.key?(:transfer_options) end end # Request passed to UpdateTransferJob. class UpdateTransferJobRequest include Google::Apis::Core::Hashable # Required. The ID of the Google Cloud Platform Console project that owns the # job. # Corresponds to the JSON property `projectId` # @return [String] attr_accessor :project_id # This resource represents the configuration of a transfer job that runs # periodically. # Corresponds to the JSON property `transferJob` # @return [Google::Apis::StoragetransferV1::TransferJob] attr_accessor :transfer_job # The field mask of the fields in `transferJob` that are to be updated in this # request. Fields in `transferJob` that can be updated are: description, # transfer_spec, notification_config, and status. To update the `transfer_spec` # of the job, a complete transfer specification must be provided. An incomplete # specification missing any required fields will be rejected with the error # INVALID_ARGUMENT. # Corresponds to the JSON property `updateTransferJobFieldMask` # @return [String] attr_accessor :update_transfer_job_field_mask def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @project_id = args[:project_id] if args.key?(:project_id) @transfer_job = args[:transfer_job] if args.key?(:transfer_job) @update_transfer_job_field_mask = args[:update_transfer_job_field_mask] if args.key?(:update_transfer_job_field_mask) end end end end end