# Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. require 'date' require 'google/apis/core/base_service' require 'google/apis/core/json_representation' require 'google/apis/core/hashable' require 'google/apis/errors' module Google module Apis module RemotebuildexecutionV2 # An `Action` captures all the information about an execution which is required # to reproduce it. `Action`s are the core component of the [Execution] service. # A single `Action` represents a repeatable action that can be performed by the # execution service. `Action`s can be succinctly identified by the digest of # their wire format encoding and, once an `Action` has been executed, will be # cached in the action cache. Future requests can then use the cached result # rather than needing to run afresh. When a server completes execution of an # Action, it MAY choose to cache the result in the ActionCache unless ` # do_not_cache` is `true`. Clients SHOULD expect the server to do so. By default, # future calls to Execute the same `Action` will also serve their results from # the cache. Clients must take care to understand the caching behaviour. Ideally, # all `Action`s will be reproducible so that serving a result from cache is # always desirable and correct. class BuildBazelRemoteExecutionV2Action include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `commandDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :command_digest # If true, then the `Action`'s result cannot be cached, and in-flight requests # for the same `Action` may not be merged. # Corresponds to the JSON property `doNotCache` # @return [Boolean] attr_accessor :do_not_cache alias_method :do_not_cache?, :do_not_cache # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `inputRootDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :input_root_digest # List of required supported NodeProperty keys. In order to ensure that # equivalent `Action`s always hash to the same value, the supported node # properties MUST be lexicographically sorted by name. Sorting of strings is # done by code point, equivalently, by the UTF-8 bytes. The interpretation of # these properties is server-dependent. If a property is not recognized by the # server, the server will return an `INVALID_ARGUMENT` error. # Corresponds to the JSON property `outputNodeProperties` # @return [Array] attr_accessor :output_node_properties # A timeout after which the execution should be killed. If the timeout is absent, # then the client is specifying that the execution should continue as long as # the server will let it. The server SHOULD impose a timeout if the client does # not specify one, however, if the client does specify a timeout that is longer # than the server's maximum timeout, the server MUST reject the request. The # timeout is a part of the Action message, and therefore two `Actions` with # different timeouts are different, even if they are otherwise identical. This # is because, if they were not, running an `Action` with a lower timeout than is # required might result in a cache hit from an execution run with a longer # timeout, hiding the fact that the timeout is too short. By encoding it # directly in the `Action`, a lower timeout will result in a cache miss and the # execution timeout will fail immediately, rather than whenever the cache entry # gets evicted. # Corresponds to the JSON property `timeout` # @return [String] attr_accessor :timeout def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @command_digest = args[:command_digest] if args.key?(:command_digest) @do_not_cache = args[:do_not_cache] if args.key?(:do_not_cache) @input_root_digest = args[:input_root_digest] if args.key?(:input_root_digest) @output_node_properties = args[:output_node_properties] if args.key?(:output_node_properties) @timeout = args[:timeout] if args.key?(:timeout) end end # Describes the server/instance capabilities for updating the action cache. class BuildBazelRemoteExecutionV2ActionCacheUpdateCapabilities include Google::Apis::Core::Hashable # # Corresponds to the JSON property `updateEnabled` # @return [Boolean] attr_accessor :update_enabled alias_method :update_enabled?, :update_enabled def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @update_enabled = args[:update_enabled] if args.key?(:update_enabled) end end # An ActionResult represents the result of an Action being run. class BuildBazelRemoteExecutionV2ActionResult include Google::Apis::Core::Hashable # ExecutedActionMetadata contains details about a completed execution. # Corresponds to the JSON property `executionMetadata` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutedActionMetadata] attr_accessor :execution_metadata # The exit code of the command. # Corresponds to the JSON property `exitCode` # @return [Fixnum] attr_accessor :exit_code # The output directories of the action. For each output directory requested in # the `output_directories` or `output_paths` field of the Action, if the # corresponding directory existed after the action completed, a single entry # will be present in the output list, which will contain the digest of a Tree # message containing the directory tree, and the path equal exactly to the # corresponding Action output_directories member. As an example, suppose the # Action had an output directory `a/b/dir` and the execution produced the # following contents in `a/b/dir`: a file named `bar` and a directory named `foo` # with an executable file named `baz`. Then, output_directory will contain ( # hashes shortened for readability): ```json // OutputDirectory proto: ` path: " # a/b/dir" tree_digest: ` hash: "4a73bc9d03...", size: 55 ` ` // Tree proto with # hash "4a73bc9d03..." and size 55: ` root: ` files: [ ` name: "bar", digest: ` # hash: "4a73bc9d03...", size: 65534 ` ` ], directories: [ ` name: "foo", digest: # ` hash: "4cf2eda940...", size: 43 ` ` ] ` children : ` // (Directory proto # with hash "4cf2eda940..." and size 43) files: [ ` name: "baz", digest: ` hash: # "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ` ``` If an output # of the same name as listed in `output_files` of the Command was found in ` # output_directories`, but was not a directory, the server will return a # FAILED_PRECONDITION. # Corresponds to the JSON property `outputDirectories` # @return [Array] attr_accessor :output_directories # The output directories of the action that are symbolic links to other # directories. Those may be links to other output directories, or input # directories, or even absolute paths outside of the working directory, if the # server supports SymlinkAbsolutePathStrategy.ALLOWED. For each output directory # requested in the `output_directories` field of the Action, if the directory # existed after the action completed, a single entry will be present either in # this field, or in the `output_directories` field, if the directory was not a # symbolic link. If an output of the same name was found, but was a symbolic # link to a file instead of a directory, the server will return a # FAILED_PRECONDITION. If the action does not produce the requested output, then # that output will be omitted from the list. The server is free to arrange the # output list as desired; clients MUST NOT assume that the output list is sorted. # DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API # should still populate this field in addition to `output_symlinks`. # Corresponds to the JSON property `outputDirectorySymlinks` # @return [Array] attr_accessor :output_directory_symlinks # The output files of the action that are symbolic links to other files. Those # may be links to other output files, or input files, or even absolute paths # outside of the working directory, if the server supports # SymlinkAbsolutePathStrategy.ALLOWED. For each output file requested in the ` # output_files` or `output_paths` field of the Action, if the corresponding file # existed after the action completed, a single entry will be present either in # this field, or in the `output_files` field, if the file was not a symbolic # link. If an output symbolic link of the same name as listed in `output_files` # of the Command was found, but its target type was not a regular file, the # server will return a FAILED_PRECONDITION. If the action does not produce the # requested output, then that output will be omitted from the list. The server # is free to arrange the output list as desired; clients MUST NOT assume that # the output list is sorted. DEPRECATED as of v2.1. Servers that wish to be # compatible with v2.0 API should still populate this field in addition to ` # output_symlinks`. # Corresponds to the JSON property `outputFileSymlinks` # @return [Array] attr_accessor :output_file_symlinks # The output files of the action. For each output file requested in the ` # output_files` or `output_paths` field of the Action, if the corresponding file # existed after the action completed, a single entry will be present either in # this field, or the `output_file_symlinks` field if the file was a symbolic # link to another file (`output_symlinks` field after v2.1). If an output listed # in `output_files` was found, but was a directory rather than a regular file, # the server will return a FAILED_PRECONDITION. If the action does not produce # the requested output, then that output will be omitted from the list. The # server is free to arrange the output list as desired; clients MUST NOT assume # that the output list is sorted. # Corresponds to the JSON property `outputFiles` # @return [Array] attr_accessor :output_files # New in v2.1: this field will only be populated if the command `output_paths` # field was used, and not the pre v2.1 `output_files` or `output_directories` # fields. The output paths of the action that are symbolic links to other paths. # Those may be links to other outputs, or inputs, or even absolute paths outside # of the working directory, if the server supports SymlinkAbsolutePathStrategy. # ALLOWED. A single entry for each output requested in `output_paths` field of # the Action, if the corresponding path existed after the action completed and # was a symbolic link. If the action does not produce a requested output, then # that output will be omitted from the list. The server is free to arrange the # output list as desired; clients MUST NOT assume that the output list is sorted. # Corresponds to the JSON property `outputSymlinks` # @return [Array] attr_accessor :output_symlinks # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `stderrDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :stderr_digest # The standard error buffer of the action. The server SHOULD NOT inline stderr # unless requested by the client in the GetActionResultRequest message. The # server MAY omit inlining, even if requested, and MUST do so if inlining would # cause the response to exceed message size limits. # Corresponds to the JSON property `stderrRaw` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :stderr_raw # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `stdoutDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :stdout_digest # The standard output buffer of the action. The server SHOULD NOT inline stdout # unless requested by the client in the GetActionResultRequest message. The # server MAY omit inlining, even if requested, and MUST do so if inlining would # cause the response to exceed message size limits. # Corresponds to the JSON property `stdoutRaw` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :stdout_raw def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @execution_metadata = args[:execution_metadata] if args.key?(:execution_metadata) @exit_code = args[:exit_code] if args.key?(:exit_code) @output_directories = args[:output_directories] if args.key?(:output_directories) @output_directory_symlinks = args[:output_directory_symlinks] if args.key?(:output_directory_symlinks) @output_file_symlinks = args[:output_file_symlinks] if args.key?(:output_file_symlinks) @output_files = args[:output_files] if args.key?(:output_files) @output_symlinks = args[:output_symlinks] if args.key?(:output_symlinks) @stderr_digest = args[:stderr_digest] if args.key?(:stderr_digest) @stderr_raw = args[:stderr_raw] if args.key?(:stderr_raw) @stdout_digest = args[:stdout_digest] if args.key?(:stdout_digest) @stdout_raw = args[:stdout_raw] if args.key?(:stdout_raw) end end # A request message for ContentAddressableStorage.BatchReadBlobs. class BuildBazelRemoteExecutionV2BatchReadBlobsRequest include Google::Apis::Core::Hashable # The individual blob digests. # Corresponds to the JSON property `digests` # @return [Array] attr_accessor :digests def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digests = args[:digests] if args.key?(:digests) end end # A response message for ContentAddressableStorage.BatchReadBlobs. class BuildBazelRemoteExecutionV2BatchReadBlobsResponse include Google::Apis::Core::Hashable # The responses to the requests. # Corresponds to the JSON property `responses` # @return [Array] attr_accessor :responses def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @responses = args[:responses] if args.key?(:responses) end end # A response corresponding to a single blob that the client tried to download. class BuildBazelRemoteExecutionV2BatchReadBlobsResponseResponse include Google::Apis::Core::Hashable # The raw binary data. # Corresponds to the JSON property `data` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :data # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @data = args[:data] if args.key?(:data) @digest = args[:digest] if args.key?(:digest) @status = args[:status] if args.key?(:status) end end # A request message for ContentAddressableStorage.BatchUpdateBlobs. class BuildBazelRemoteExecutionV2BatchUpdateBlobsRequest include Google::Apis::Core::Hashable # The individual upload requests. # Corresponds to the JSON property `requests` # @return [Array] attr_accessor :requests def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @requests = args[:requests] if args.key?(:requests) end end # A request corresponding to a single blob that the client wants to upload. class BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest include Google::Apis::Core::Hashable # The raw binary data. # Corresponds to the JSON property `data` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :data # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @data = args[:data] if args.key?(:data) @digest = args[:digest] if args.key?(:digest) end end # A response message for ContentAddressableStorage.BatchUpdateBlobs. class BuildBazelRemoteExecutionV2BatchUpdateBlobsResponse include Google::Apis::Core::Hashable # The responses to the requests. # Corresponds to the JSON property `responses` # @return [Array] attr_accessor :responses def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @responses = args[:responses] if args.key?(:responses) end end # A response corresponding to a single blob that the client tried to upload. class BuildBazelRemoteExecutionV2BatchUpdateBlobsResponseResponse include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest = args[:digest] if args.key?(:digest) @status = args[:status] if args.key?(:status) end end # Capabilities of the remote cache system. class BuildBazelRemoteExecutionV2CacheCapabilities include Google::Apis::Core::Hashable # Describes the server/instance capabilities for updating the action cache. # Corresponds to the JSON property `actionCacheUpdateCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ActionCacheUpdateCapabilities] attr_accessor :action_cache_update_capabilities # Allowed values for priority in ResultsCachePolicy Used for querying both cache # and execution valid priority ranges. # Corresponds to the JSON property `cachePriorityCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PriorityCapabilities] attr_accessor :cache_priority_capabilities # All the digest functions supported by the remote cache. Remote cache may # support multiple digest functions simultaneously. # Corresponds to the JSON property `digestFunction` # @return [Array] attr_accessor :digest_function # Maximum total size of blobs to be uploaded/downloaded using batch methods. A # value of 0 means no limit is set, although in practice there will always be a # message size limitation of the protocol in use, e.g. GRPC. # Corresponds to the JSON property `maxBatchTotalSizeBytes` # @return [Fixnum] attr_accessor :max_batch_total_size_bytes # Whether absolute symlink targets are supported. # Corresponds to the JSON property `symlinkAbsolutePathStrategy` # @return [String] attr_accessor :symlink_absolute_path_strategy def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @action_cache_update_capabilities = args[:action_cache_update_capabilities] if args.key?(:action_cache_update_capabilities) @cache_priority_capabilities = args[:cache_priority_capabilities] if args.key?(:cache_priority_capabilities) @digest_function = args[:digest_function] if args.key?(:digest_function) @max_batch_total_size_bytes = args[:max_batch_total_size_bytes] if args.key?(:max_batch_total_size_bytes) @symlink_absolute_path_strategy = args[:symlink_absolute_path_strategy] if args.key?(:symlink_absolute_path_strategy) end end # A `Command` is the actual command executed by a worker running an Action and # specifications of its environment. Except as otherwise required, the # environment (such as which system libraries or binaries are available, and # what filesystems are mounted where) is defined by and specific to the # implementation of the remote execution API. class BuildBazelRemoteExecutionV2Command include Google::Apis::Core::Hashable # The arguments to the command. The first argument must be the path to the # executable, which must be either a relative path, in which case it is # evaluated with respect to the input root, or an absolute path. # Corresponds to the JSON property `arguments` # @return [Array] attr_accessor :arguments # The environment variables to set when running the program. The worker may # provide its own default environment variables; these defaults can be # overridden using this field. Additional variables can also be specified. In # order to ensure that equivalent Commands always hash to the same value, the # environment variables MUST be lexicographically sorted by name. Sorting of # strings is done by code point, equivalently, by the UTF-8 bytes. # Corresponds to the JSON property `environmentVariables` # @return [Array] attr_accessor :environment_variables # A list of the output directories that the client expects to retrieve from the # action. Only the listed directories will be returned (an entire directory # structure will be returned as a Tree message digest, see OutputDirectory), as # well as files listed in `output_files`. Other files or directories that may be # created during command execution are discarded. The paths are relative to the # working directory of the action execution. The paths are specified using a # single forward slash (`/`) as a path separator, even if the execution platform # natively uses a different separator. The path MUST NOT include a trailing # slash, nor a leading slash, being a relative path. The special value of empty # string is allowed, although not recommended, and can be used to capture the # entire working directory tree, including inputs. In order to ensure consistent # hashing of the same Action, the output paths MUST be sorted lexicographically # by code point (or, equivalently, by UTF-8 bytes). An output directory cannot # be duplicated or have the same path as any of the listed output files. An # output directory is allowed to be a parent of another output directory. # Directories leading up to the output directories (but not the output # directories themselves) are created by the worker prior to execution, even if # they are not explicitly part of the input root. DEPRECATED since 2.1: Use ` # output_paths` instead. # Corresponds to the JSON property `outputDirectories` # @return [Array] attr_accessor :output_directories # A list of the output files that the client expects to retrieve from the action. # Only the listed files, as well as directories listed in `output_directories`, # will be returned to the client as output. Other files or directories that may # be created during command execution are discarded. The paths are relative to # the working directory of the action execution. The paths are specified using a # single forward slash (`/`) as a path separator, even if the execution platform # natively uses a different separator. The path MUST NOT include a trailing # slash, nor a leading slash, being a relative path. In order to ensure # consistent hashing of the same Action, the output paths MUST be sorted # lexicographically by code point (or, equivalently, by UTF-8 bytes). An output # file cannot be duplicated, be a parent of another output file, or have the # same path as any of the listed output directories. Directories leading up to # the output files are created by the worker prior to execution, even if they # are not explicitly part of the input root. DEPRECATED since v2.1: Use ` # output_paths` instead. # Corresponds to the JSON property `outputFiles` # @return [Array] attr_accessor :output_files # A list of the output paths that the client expects to retrieve from the action. # Only the listed paths will be returned to the client as output. The type of # the output (file or directory) is not specified, and will be determined by the # server after action execution. If the resulting path is a file, it will be # returned in an OutputFile) typed field. If the path is a directory, the entire # directory structure will be returned as a Tree message digest, see # OutputDirectory) Other files or directories that may be created during command # execution are discarded. The paths are relative to the working directory of # the action execution. The paths are specified using a single forward slash (`/` # ) as a path separator, even if the execution platform natively uses a # different separator. The path MUST NOT include a trailing slash, nor a leading # slash, being a relative path. In order to ensure consistent hashing of the # same Action, the output paths MUST be deduplicated and sorted # lexicographically by code point (or, equivalently, by UTF-8 bytes). # Directories leading up to the output paths are created by the worker prior to # execution, even if they are not explicitly part of the input root. New in v2.1: # this field supersedes the DEPRECATED `output_files` and `output_directories` # fields. If `output_paths` is used, `output_files` and `output_directories` # will be ignored! # Corresponds to the JSON property `outputPaths` # @return [Array] attr_accessor :output_paths # A `Platform` is a set of requirements, such as hardware, operating system, or # compiler toolchain, for an Action's execution environment. A `Platform` is # represented as a series of key-value pairs representing the properties that # are required of the platform. # Corresponds to the JSON property `platform` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Platform] attr_accessor :platform # The working directory, relative to the input root, for the command to run in. # It must be a directory which exists in the input tree. If it is left empty, # then the action is run in the input root. # Corresponds to the JSON property `workingDirectory` # @return [String] attr_accessor :working_directory def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @arguments = args[:arguments] if args.key?(:arguments) @environment_variables = args[:environment_variables] if args.key?(:environment_variables) @output_directories = args[:output_directories] if args.key?(:output_directories) @output_files = args[:output_files] if args.key?(:output_files) @output_paths = args[:output_paths] if args.key?(:output_paths) @platform = args[:platform] if args.key?(:platform) @working_directory = args[:working_directory] if args.key?(:working_directory) end end # An `EnvironmentVariable` is one variable to set in the running program's # environment. class BuildBazelRemoteExecutionV2CommandEnvironmentVariable include Google::Apis::Core::Hashable # The variable name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The variable value. # Corresponds to the JSON property `value` # @return [String] attr_accessor :value def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @value = args[:value] if args.key?(:value) end end # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. class BuildBazelRemoteExecutionV2Digest include Google::Apis::Core::Hashable # The hash. In the case of SHA-256, it will always be a lowercase hex string # exactly 64 characters long. # Corresponds to the JSON property `hash` # @return [String] attr_accessor :hash_prop # The size of the blob, in bytes. # Corresponds to the JSON property `sizeBytes` # @return [Fixnum] attr_accessor :size_bytes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @hash_prop = args[:hash_prop] if args.key?(:hash_prop) @size_bytes = args[:size_bytes] if args.key?(:size_bytes) end end # A `Directory` represents a directory node in a file tree, containing zero or # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains # its name in the directory, either the digest of its content (either a file # blob or a `Directory` proto) or a symlink target, as well as possibly some # metadata about the file or directory. In order to ensure that two equivalent # directory trees hash to the same value, the following restrictions MUST be # obeyed when constructing a a `Directory`: * Every child in the directory must # have a path of exactly one segment. Multiple levels of directory hierarchy may # not be collapsed. * Each child in the directory must have a unique path # segment (file name). Note that while the API itself is case-sensitive, the # environment where the Action is executed may or may not be case-sensitive. # That is, it is legal to call the API with a Directory that has both "Foo" and " # foo" as children, but the Action may be rejected by the remote system upon # execution. * The files, directories and symlinks in the directory must each be # sorted in lexicographical order by path. The path strings must be sorted by # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files, # directories, and symlinks must be sorted in lexicographical order by property # name. A `Directory` that obeys the restrictions is said to be in canonical # form. As an example, the following could be used for a file named `bar` and a # directory named `foo` with an executable file named `baz` (hashes shortened # for readability): ```json // (Directory proto) ` files: [ ` name: "bar", # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": " # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: " # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: ` # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ``` class BuildBazelRemoteExecutionV2Directory include Google::Apis::Core::Hashable # The subdirectories in the directory. # Corresponds to the JSON property `directories` # @return [Array] attr_accessor :directories # The files in the directory. # Corresponds to the JSON property `files` # @return [Array] attr_accessor :files # The node properties of the Directory. # Corresponds to the JSON property `nodeProperties` # @return [Array] attr_accessor :node_properties # The symlinks in the directory. # Corresponds to the JSON property `symlinks` # @return [Array] attr_accessor :symlinks def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @directories = args[:directories] if args.key?(:directories) @files = args[:files] if args.key?(:files) @node_properties = args[:node_properties] if args.key?(:node_properties) @symlinks = args[:symlinks] if args.key?(:symlinks) end end # A `DirectoryNode` represents a child of a Directory which is itself a ` # Directory` and its associated metadata. class BuildBazelRemoteExecutionV2DirectoryNode include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # The name of the directory. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest = args[:digest] if args.key?(:digest) @name = args[:name] if args.key?(:name) end end # Metadata about an ongoing execution, which will be contained in the metadata # field of the Operation. class BuildBazelRemoteExecutionV2ExecuteOperationMetadata include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `actionDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :action_digest # The current stage of execution. # Corresponds to the JSON property `stage` # @return [String] attr_accessor :stage # If set, the client can use this name with ByteStream.Read to stream the # standard error. # Corresponds to the JSON property `stderrStreamName` # @return [String] attr_accessor :stderr_stream_name # If set, the client can use this name with ByteStream.Read to stream the # standard output. # Corresponds to the JSON property `stdoutStreamName` # @return [String] attr_accessor :stdout_stream_name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @action_digest = args[:action_digest] if args.key?(:action_digest) @stage = args[:stage] if args.key?(:stage) @stderr_stream_name = args[:stderr_stream_name] if args.key?(:stderr_stream_name) @stdout_stream_name = args[:stdout_stream_name] if args.key?(:stdout_stream_name) end end # A request message for Execution.Execute. class BuildBazelRemoteExecutionV2ExecuteRequest include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `actionDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :action_digest # An `ExecutionPolicy` can be used to control the scheduling of the action. # Corresponds to the JSON property `executionPolicy` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutionPolicy] attr_accessor :execution_policy # A `ResultsCachePolicy` is used for fine-grained control over how action # outputs are stored in the CAS and Action Cache. # Corresponds to the JSON property `resultsCachePolicy` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ResultsCachePolicy] attr_accessor :results_cache_policy # If true, the action will be executed even if its result is already present in # the ActionCache. The execution is still allowed to be merged with other in- # flight executions of the same action, however - semantically, the service MUST # only guarantee that the results of an execution with this field set were not # visible before the corresponding execution request was sent. Note that actions # from execution requests setting this field set are still eligible to be # entered into the action cache upon completion, and services SHOULD overwrite # any existing entries that may exist. This allows skip_cache_lookup requests to # be used as a mechanism for replacing action cache entries that reference # outputs no longer available or that are poisoned in any way. If false, the # result may be served from the action cache. # Corresponds to the JSON property `skipCacheLookup` # @return [Boolean] attr_accessor :skip_cache_lookup alias_method :skip_cache_lookup?, :skip_cache_lookup def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @action_digest = args[:action_digest] if args.key?(:action_digest) @execution_policy = args[:execution_policy] if args.key?(:execution_policy) @results_cache_policy = args[:results_cache_policy] if args.key?(:results_cache_policy) @skip_cache_lookup = args[:skip_cache_lookup] if args.key?(:skip_cache_lookup) end end # The response message for Execution.Execute, which will be contained in the # response field of the Operation. class BuildBazelRemoteExecutionV2ExecuteResponse include Google::Apis::Core::Hashable # True if the result was served from cache, false if it was executed. # Corresponds to the JSON property `cachedResult` # @return [Boolean] attr_accessor :cached_result alias_method :cached_result?, :cached_result # Freeform informational message with details on the execution of the action # that may be displayed to the user upon failure or when requested explicitly. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message # An ActionResult represents the result of an Action being run. # Corresponds to the JSON property `result` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ActionResult] attr_accessor :result # An optional list of additional log outputs the server wishes to provide. A # server can use this to return execution-specific logs however it wishes. This # is intended primarily to make it easier for users to debug issues that may be # outside of the actual job execution, such as by identifying the worker # executing the action or by providing logs from the worker's setup phase. The # keys SHOULD be human readable so that a client can display them to a user. # Corresponds to the JSON property `serverLogs` # @return [Hash] attr_accessor :server_logs # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cached_result = args[:cached_result] if args.key?(:cached_result) @message = args[:message] if args.key?(:message) @result = args[:result] if args.key?(:result) @server_logs = args[:server_logs] if args.key?(:server_logs) @status = args[:status] if args.key?(:status) end end # ExecutedActionMetadata contains details about a completed execution. class BuildBazelRemoteExecutionV2ExecutedActionMetadata include Google::Apis::Core::Hashable # When the worker completed executing the action command. # Corresponds to the JSON property `executionCompletedTimestamp` # @return [String] attr_accessor :execution_completed_timestamp # When the worker started executing the action command. # Corresponds to the JSON property `executionStartTimestamp` # @return [String] attr_accessor :execution_start_timestamp # When the worker finished fetching action inputs. # Corresponds to the JSON property `inputFetchCompletedTimestamp` # @return [String] attr_accessor :input_fetch_completed_timestamp # When the worker started fetching action inputs. # Corresponds to the JSON property `inputFetchStartTimestamp` # @return [String] attr_accessor :input_fetch_start_timestamp # When the worker finished uploading action outputs. # Corresponds to the JSON property `outputUploadCompletedTimestamp` # @return [String] attr_accessor :output_upload_completed_timestamp # When the worker started uploading action outputs. # Corresponds to the JSON property `outputUploadStartTimestamp` # @return [String] attr_accessor :output_upload_start_timestamp # When was the action added to the queue. # Corresponds to the JSON property `queuedTimestamp` # @return [String] attr_accessor :queued_timestamp # The name of the worker which ran the execution. # Corresponds to the JSON property `worker` # @return [String] attr_accessor :worker # When the worker completed the action, including all stages. # Corresponds to the JSON property `workerCompletedTimestamp` # @return [String] attr_accessor :worker_completed_timestamp # When the worker received the action. # Corresponds to the JSON property `workerStartTimestamp` # @return [String] attr_accessor :worker_start_timestamp def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @execution_completed_timestamp = args[:execution_completed_timestamp] if args.key?(:execution_completed_timestamp) @execution_start_timestamp = args[:execution_start_timestamp] if args.key?(:execution_start_timestamp) @input_fetch_completed_timestamp = args[:input_fetch_completed_timestamp] if args.key?(:input_fetch_completed_timestamp) @input_fetch_start_timestamp = args[:input_fetch_start_timestamp] if args.key?(:input_fetch_start_timestamp) @output_upload_completed_timestamp = args[:output_upload_completed_timestamp] if args.key?(:output_upload_completed_timestamp) @output_upload_start_timestamp = args[:output_upload_start_timestamp] if args.key?(:output_upload_start_timestamp) @queued_timestamp = args[:queued_timestamp] if args.key?(:queued_timestamp) @worker = args[:worker] if args.key?(:worker) @worker_completed_timestamp = args[:worker_completed_timestamp] if args.key?(:worker_completed_timestamp) @worker_start_timestamp = args[:worker_start_timestamp] if args.key?(:worker_start_timestamp) end end # Capabilities of the remote execution system. class BuildBazelRemoteExecutionV2ExecutionCapabilities include Google::Apis::Core::Hashable # Remote execution may only support a single digest function. # Corresponds to the JSON property `digestFunction` # @return [String] attr_accessor :digest_function # Whether remote execution is enabled for the particular server/instance. # Corresponds to the JSON property `execEnabled` # @return [Boolean] attr_accessor :exec_enabled alias_method :exec_enabled?, :exec_enabled # Allowed values for priority in ResultsCachePolicy Used for querying both cache # and execution valid priority ranges. # Corresponds to the JSON property `executionPriorityCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PriorityCapabilities] attr_accessor :execution_priority_capabilities # Supported node properties. # Corresponds to the JSON property `supportedNodeProperties` # @return [Array] attr_accessor :supported_node_properties def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest_function = args[:digest_function] if args.key?(:digest_function) @exec_enabled = args[:exec_enabled] if args.key?(:exec_enabled) @execution_priority_capabilities = args[:execution_priority_capabilities] if args.key?(:execution_priority_capabilities) @supported_node_properties = args[:supported_node_properties] if args.key?(:supported_node_properties) end end # An `ExecutionPolicy` can be used to control the scheduling of the action. class BuildBazelRemoteExecutionV2ExecutionPolicy include Google::Apis::Core::Hashable # The priority (relative importance) of this action. Generally, a lower value # means that the action should be run sooner than actions having a greater # priority value, but the interpretation of a given value is server- dependent. # A priority of 0 means the *default* priority. Priorities may be positive or # negative, and such actions should run later or sooner than actions having the # default priority, respectively. The particular semantics of this field is up # to the server. In particular, every server will have their own supported range # of priorities, and will decide how these map into scheduling policy. # Corresponds to the JSON property `priority` # @return [Fixnum] attr_accessor :priority def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @priority = args[:priority] if args.key?(:priority) end end # A `FileNode` represents a single file and associated metadata. class BuildBazelRemoteExecutionV2FileNode include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # True if file is executable, false otherwise. # Corresponds to the JSON property `isExecutable` # @return [Boolean] attr_accessor :is_executable alias_method :is_executable?, :is_executable # The name of the file. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The node properties of the FileNode. # Corresponds to the JSON property `nodeProperties` # @return [Array] attr_accessor :node_properties def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest = args[:digest] if args.key?(:digest) @is_executable = args[:is_executable] if args.key?(:is_executable) @name = args[:name] if args.key?(:name) @node_properties = args[:node_properties] if args.key?(:node_properties) end end # A request message for ContentAddressableStorage.FindMissingBlobs. class BuildBazelRemoteExecutionV2FindMissingBlobsRequest include Google::Apis::Core::Hashable # A list of the blobs to check. # Corresponds to the JSON property `blobDigests` # @return [Array] attr_accessor :blob_digests def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @blob_digests = args[:blob_digests] if args.key?(:blob_digests) end end # A response message for ContentAddressableStorage.FindMissingBlobs. class BuildBazelRemoteExecutionV2FindMissingBlobsResponse include Google::Apis::Core::Hashable # A list of the blobs requested *not* present in the storage. # Corresponds to the JSON property `missingBlobDigests` # @return [Array] attr_accessor :missing_blob_digests def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @missing_blob_digests = args[:missing_blob_digests] if args.key?(:missing_blob_digests) end end # A response message for ContentAddressableStorage.GetTree. class BuildBazelRemoteExecutionV2GetTreeResponse include Google::Apis::Core::Hashable # The directories descended from the requested root. # Corresponds to the JSON property `directories` # @return [Array] attr_accessor :directories # If present, signifies that there are more results which the client can # retrieve by passing this as the page_token in a subsequent request. If empty, # signifies that this is the last page of results. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @directories = args[:directories] if args.key?(:directories) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # A `LogFile` is a log stored in the CAS. class BuildBazelRemoteExecutionV2LogFile include Google::Apis::Core::Hashable # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # This is a hint as to the purpose of the log, and is set to true if the log is # human-readable text that can be usefully displayed to a user, and false # otherwise. For instance, if a command-line client wishes to print the server # logs to the terminal for a failed action, this allows it to avoid displaying a # binary file. # Corresponds to the JSON property `humanReadable` # @return [Boolean] attr_accessor :human_readable alias_method :human_readable?, :human_readable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest = args[:digest] if args.key?(:digest) @human_readable = args[:human_readable] if args.key?(:human_readable) end end # A single property for FileNodes, DirectoryNodes, and SymlinkNodes. The server # is responsible for specifying the property `name`s that it accepts. If # permitted by the server, the same `name` may occur multiple times. class BuildBazelRemoteExecutionV2NodeProperty include Google::Apis::Core::Hashable # The property name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The property value. # Corresponds to the JSON property `value` # @return [String] attr_accessor :value def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @value = args[:value] if args.key?(:value) end end # An `OutputDirectory` is the output in an `ActionResult` corresponding to a # directory's full contents rather than a single file. class BuildBazelRemoteExecutionV2OutputDirectory include Google::Apis::Core::Hashable # The full path of the directory relative to the working directory. The path # separator is a forward slash `/`. Since this is a relative path, it MUST NOT # begin with a leading forward slash. The empty string value is allowed, and it # denotes the entire working directory. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `treeDigest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :tree_digest def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @path = args[:path] if args.key?(:path) @tree_digest = args[:tree_digest] if args.key?(:tree_digest) end end # An `OutputFile` is similar to a FileNode, but it is used as an output in an ` # ActionResult`. It allows a full file path rather than only a name. class BuildBazelRemoteExecutionV2OutputFile include Google::Apis::Core::Hashable # The contents of the file if inlining was requested. The server SHOULD NOT # inline file contents unless requested by the client in the # GetActionResultRequest message. The server MAY omit inlining, even if # requested, and MUST do so if inlining would cause the response to exceed # message size limits. # Corresponds to the JSON property `contents` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :contents # A content digest. A digest for a given blob consists of the size of the blob # and its hash. The hash algorithm to use is defined by the server. The size is # considered to be an integral part of the digest and cannot be separated. That # is, even if the `hash` field is correctly specified but `size_bytes` is not, # the server MUST reject the request. The reason for including the size in the # digest is as follows: in a great many cases, the server needs to know the size # of the blob it is about to work with prior to starting an operation with it, # such as flattening Merkle tree structures or streaming it to a worker. # Technically, the server could implement a separate metadata store, but this # results in a significantly more complicated implementation as opposed to # having the client specify the size up-front (or storing the size along with # the digest in every message where digests are embedded). This does mean that # the API leaks some implementation details of (what we consider to be) a # reasonable server implementation, but we consider this to be a worthwhile # tradeoff. When a `Digest` is used to refer to a proto message, it always # refers to the message in binary encoded form. To ensure consistent hashing, # clients and servers MUST ensure that they serialize messages according to the # following rules, even if there are alternate valid encodings for the same # message: * Fields are serialized in tag order. * There are no unknown fields. * # There are no duplicate fields. * Fields are serialized according to the # default semantics for their type. Most protocol buffer implementations will # always follow these rules when serializing, but care should be taken to avoid # shortcuts. For instance, concatenating two messages to merge them may produce # duplicate fields. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest] attr_accessor :digest # True if file is executable, false otherwise. # Corresponds to the JSON property `isExecutable` # @return [Boolean] attr_accessor :is_executable alias_method :is_executable?, :is_executable # The supported node properties of the OutputFile, if requested by the Action. # Corresponds to the JSON property `nodeProperties` # @return [Array] attr_accessor :node_properties # The full path of the file relative to the working directory, including the # filename. The path separator is a forward slash `/`. Since this is a relative # path, it MUST NOT begin with a leading forward slash. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @contents = args[:contents] if args.key?(:contents) @digest = args[:digest] if args.key?(:digest) @is_executable = args[:is_executable] if args.key?(:is_executable) @node_properties = args[:node_properties] if args.key?(:node_properties) @path = args[:path] if args.key?(:path) end end # An `OutputSymlink` is similar to a Symlink, but it is used as an output in an ` # ActionResult`. `OutputSymlink` is binary-compatible with `SymlinkNode`. class BuildBazelRemoteExecutionV2OutputSymlink include Google::Apis::Core::Hashable # The supported node properties of the OutputSymlink, if requested by the Action. # Corresponds to the JSON property `nodeProperties` # @return [Array] attr_accessor :node_properties # The full path of the symlink relative to the working directory, including the # filename. The path separator is a forward slash `/`. Since this is a relative # path, it MUST NOT begin with a leading forward slash. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path # The target path of the symlink. The path separator is a forward slash `/`. The # target path can be relative to the parent directory of the symlink or it can # be an absolute path starting with `/`. Support for absolute paths can be # checked using the Capabilities API. The canonical form forbids the substrings ` # /./` and `//` in the target path. `..` components are allowed anywhere in the # target path. # Corresponds to the JSON property `target` # @return [String] attr_accessor :target def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @node_properties = args[:node_properties] if args.key?(:node_properties) @path = args[:path] if args.key?(:path) @target = args[:target] if args.key?(:target) end end # A `Platform` is a set of requirements, such as hardware, operating system, or # compiler toolchain, for an Action's execution environment. A `Platform` is # represented as a series of key-value pairs representing the properties that # are required of the platform. class BuildBazelRemoteExecutionV2Platform include Google::Apis::Core::Hashable # The properties that make up this platform. In order to ensure that equivalent ` # Platform`s always hash to the same value, the properties MUST be # lexicographically sorted by name, and then by value. Sorting of strings is # done by code point, equivalently, by the UTF-8 bytes. # Corresponds to the JSON property `properties` # @return [Array] attr_accessor :properties def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @properties = args[:properties] if args.key?(:properties) end end # A single property for the environment. The server is responsible for # specifying the property `name`s that it accepts. If an unknown `name` is # provided in the requirements for an Action, the server SHOULD reject the # execution request. If permitted by the server, the same `name` may occur # multiple times. The server is also responsible for specifying the # interpretation of property `value`s. For instance, a property describing how # much RAM must be available may be interpreted as allowing a worker with 16GB # to fulfill a request for 8GB, while a property describing the OS environment # on which the action must be performed may require an exact match with the # worker's OS. The server MAY use the `value` of one or more properties to # determine how it sets up the execution environment, such as by making specific # system files available to the worker. class BuildBazelRemoteExecutionV2PlatformProperty include Google::Apis::Core::Hashable # The property name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The property value. # Corresponds to the JSON property `value` # @return [String] attr_accessor :value def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @value = args[:value] if args.key?(:value) end end # Allowed values for priority in ResultsCachePolicy Used for querying both cache # and execution valid priority ranges. class BuildBazelRemoteExecutionV2PriorityCapabilities include Google::Apis::Core::Hashable # # Corresponds to the JSON property `priorities` # @return [Array] attr_accessor :priorities def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @priorities = args[:priorities] if args.key?(:priorities) end end # Supported range of priorities, including boundaries. class BuildBazelRemoteExecutionV2PriorityCapabilitiesPriorityRange include Google::Apis::Core::Hashable # # Corresponds to the JSON property `maxPriority` # @return [Fixnum] attr_accessor :max_priority # # Corresponds to the JSON property `minPriority` # @return [Fixnum] attr_accessor :min_priority def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @max_priority = args[:max_priority] if args.key?(:max_priority) @min_priority = args[:min_priority] if args.key?(:min_priority) end end # An optional Metadata to attach to any RPC request to tell the server about an # external context of the request. The server may use this for logging or other # purposes. To use it, the client attaches the header to the call using the # canonical proto serialization: * name: `build.bazel.remote.execution.v2. # requestmetadata-bin` * contents: the base64 encoded binary `RequestMetadata` # message. Note: the gRPC library serializes binary headers encoded in base 64 # by default (https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md# # requests). Therefore, if the gRPC library is used to pass/retrieve this # metadata, the user may ignore the base64 encoding and assume it is simply # serialized as a binary message. class BuildBazelRemoteExecutionV2RequestMetadata include Google::Apis::Core::Hashable # An identifier that ties multiple requests to the same action. For example, # multiple requests to the CAS, Action Cache, and Execution API are used in # order to compile foo.cc. # Corresponds to the JSON property `actionId` # @return [String] attr_accessor :action_id # An identifier to tie multiple tool invocations together. For example, runs of # foo_test, bar_test and baz_test on a post-submit of a given patch. # Corresponds to the JSON property `correlatedInvocationsId` # @return [String] attr_accessor :correlated_invocations_id # Details for the tool used to call the API. # Corresponds to the JSON property `toolDetails` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ToolDetails] attr_accessor :tool_details # An identifier that ties multiple actions together to a final result. For # example, multiple actions are required to build and run foo_test. # Corresponds to the JSON property `toolInvocationId` # @return [String] attr_accessor :tool_invocation_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @action_id = args[:action_id] if args.key?(:action_id) @correlated_invocations_id = args[:correlated_invocations_id] if args.key?(:correlated_invocations_id) @tool_details = args[:tool_details] if args.key?(:tool_details) @tool_invocation_id = args[:tool_invocation_id] if args.key?(:tool_invocation_id) end end # A `ResultsCachePolicy` is used for fine-grained control over how action # outputs are stored in the CAS and Action Cache. class BuildBazelRemoteExecutionV2ResultsCachePolicy include Google::Apis::Core::Hashable # The priority (relative importance) of this content in the overall cache. # Generally, a lower value means a longer retention time or other advantage, but # the interpretation of a given value is server-dependent. A priority of 0 means # a *default* value, decided by the server. The particular semantics of this # field is up to the server. In particular, every server will have their own # supported range of priorities, and will decide how these map into retention/ # eviction policy. # Corresponds to the JSON property `priority` # @return [Fixnum] attr_accessor :priority def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @priority = args[:priority] if args.key?(:priority) end end # A response message for Capabilities.GetCapabilities. class BuildBazelRemoteExecutionV2ServerCapabilities include Google::Apis::Core::Hashable # Capabilities of the remote cache system. # Corresponds to the JSON property `cacheCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2CacheCapabilities] attr_accessor :cache_capabilities # The full version of a given tool. # Corresponds to the JSON property `deprecatedApiVersion` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer] attr_accessor :deprecated_api_version # Capabilities of the remote execution system. # Corresponds to the JSON property `executionCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutionCapabilities] attr_accessor :execution_capabilities # The full version of a given tool. # Corresponds to the JSON property `highApiVersion` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer] attr_accessor :high_api_version # The full version of a given tool. # Corresponds to the JSON property `lowApiVersion` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer] attr_accessor :low_api_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cache_capabilities = args[:cache_capabilities] if args.key?(:cache_capabilities) @deprecated_api_version = args[:deprecated_api_version] if args.key?(:deprecated_api_version) @execution_capabilities = args[:execution_capabilities] if args.key?(:execution_capabilities) @high_api_version = args[:high_api_version] if args.key?(:high_api_version) @low_api_version = args[:low_api_version] if args.key?(:low_api_version) end end # A `SymlinkNode` represents a symbolic link. class BuildBazelRemoteExecutionV2SymlinkNode include Google::Apis::Core::Hashable # The name of the symlink. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The node properties of the SymlinkNode. # Corresponds to the JSON property `nodeProperties` # @return [Array] attr_accessor :node_properties # The target path of the symlink. The path separator is a forward slash `/`. The # target path can be relative to the parent directory of the symlink or it can # be an absolute path starting with `/`. Support for absolute paths can be # checked using the Capabilities API. The canonical form forbids the substrings ` # /./` and `//` in the target path. `..` components are allowed anywhere in the # target path. # Corresponds to the JSON property `target` # @return [String] attr_accessor :target def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @node_properties = args[:node_properties] if args.key?(:node_properties) @target = args[:target] if args.key?(:target) end end # Details for the tool used to call the API. class BuildBazelRemoteExecutionV2ToolDetails include Google::Apis::Core::Hashable # Name of the tool, e.g. bazel. # Corresponds to the JSON property `toolName` # @return [String] attr_accessor :tool_name # Version of the tool used for the request, e.g. 5.0.3. # Corresponds to the JSON property `toolVersion` # @return [String] attr_accessor :tool_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @tool_name = args[:tool_name] if args.key?(:tool_name) @tool_version = args[:tool_version] if args.key?(:tool_version) end end # A `Tree` contains all the Directory protos in a single directory Merkle tree, # compressed into one message. class BuildBazelRemoteExecutionV2Tree include Google::Apis::Core::Hashable # All the child directories: the directories referred to by the root and, # recursively, all its children. In order to reconstruct the directory tree, the # client must take the digests of each of the child directories and then build # up a tree starting from the `root`. # Corresponds to the JSON property `children` # @return [Array] attr_accessor :children # A `Directory` represents a directory node in a file tree, containing zero or # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains # its name in the directory, either the digest of its content (either a file # blob or a `Directory` proto) or a symlink target, as well as possibly some # metadata about the file or directory. In order to ensure that two equivalent # directory trees hash to the same value, the following restrictions MUST be # obeyed when constructing a a `Directory`: * Every child in the directory must # have a path of exactly one segment. Multiple levels of directory hierarchy may # not be collapsed. * Each child in the directory must have a unique path # segment (file name). Note that while the API itself is case-sensitive, the # environment where the Action is executed may or may not be case-sensitive. # That is, it is legal to call the API with a Directory that has both "Foo" and " # foo" as children, but the Action may be rejected by the remote system upon # execution. * The files, directories and symlinks in the directory must each be # sorted in lexicographical order by path. The path strings must be sorted by # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files, # directories, and symlinks must be sorted in lexicographical order by property # name. A `Directory` that obeys the restrictions is said to be in canonical # form. As an example, the following could be used for a file named `bar` and a # directory named `foo` with an executable file named `baz` (hashes shortened # for readability): ```json // (Directory proto) ` files: [ ` name: "bar", # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": " # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: " # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: ` # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ``` # Corresponds to the JSON property `root` # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Directory] attr_accessor :root def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @children = args[:children] if args.key?(:children) @root = args[:root] if args.key?(:root) end end # A request message for WaitExecution. class BuildBazelRemoteExecutionV2WaitExecutionRequest include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # The full version of a given tool. class BuildBazelSemverSemVer include Google::Apis::Core::Hashable # The major version, e.g 10 for 10.2.3. # Corresponds to the JSON property `major` # @return [Fixnum] attr_accessor :major # The minor version, e.g. 2 for 10.2.3. # Corresponds to the JSON property `minor` # @return [Fixnum] attr_accessor :minor # The patch version, e.g 3 for 10.2.3. # Corresponds to the JSON property `patch` # @return [Fixnum] attr_accessor :patch # The pre-release version. Either this field or major/minor/patch fields must be # filled. They are mutually exclusive. Pre-release versions are assumed to be # earlier than any released versions. # Corresponds to the JSON property `prerelease` # @return [String] attr_accessor :prerelease def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @major = args[:major] if args.key?(:major) @minor = args[:minor] if args.key?(:minor) @patch = args[:patch] if args.key?(:patch) @prerelease = args[:prerelease] if args.key?(:prerelease) end end # CommandDuration contains the various duration metrics tracked when a bot # performs a command. class GoogleDevtoolsRemotebuildbotCommandDurations include Google::Apis::Core::Hashable # The time spent preparing the command to be run in a Docker container (includes # pulling the Docker image, if necessary). # Corresponds to the JSON property `dockerPrep` # @return [String] attr_accessor :docker_prep # The timestamp when docker preparation begins. # Corresponds to the JSON property `dockerPrepStartTime` # @return [String] attr_accessor :docker_prep_start_time # The time spent downloading the input files and constructing the working # directory. # Corresponds to the JSON property `download` # @return [String] attr_accessor :download # The timestamp when downloading the input files begins. # Corresponds to the JSON property `downloadStartTime` # @return [String] attr_accessor :download_start_time # The timestamp when execution begins. # Corresponds to the JSON property `execStartTime` # @return [String] attr_accessor :exec_start_time # The time spent executing the command (i.e., doing useful work). # Corresponds to the JSON property `execution` # @return [String] attr_accessor :execution # The timestamp when preparation is done and bot starts downloading files. # Corresponds to the JSON property `isoPrepDone` # @return [String] attr_accessor :iso_prep_done # The time spent completing the command, in total. # Corresponds to the JSON property `overall` # @return [String] attr_accessor :overall # The time spent uploading the stdout logs. # Corresponds to the JSON property `stdout` # @return [String] attr_accessor :stdout # The time spent uploading the output files. # Corresponds to the JSON property `upload` # @return [String] attr_accessor :upload # The timestamp when uploading the output files begins. # Corresponds to the JSON property `uploadStartTime` # @return [String] attr_accessor :upload_start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @docker_prep = args[:docker_prep] if args.key?(:docker_prep) @docker_prep_start_time = args[:docker_prep_start_time] if args.key?(:docker_prep_start_time) @download = args[:download] if args.key?(:download) @download_start_time = args[:download_start_time] if args.key?(:download_start_time) @exec_start_time = args[:exec_start_time] if args.key?(:exec_start_time) @execution = args[:execution] if args.key?(:execution) @iso_prep_done = args[:iso_prep_done] if args.key?(:iso_prep_done) @overall = args[:overall] if args.key?(:overall) @stdout = args[:stdout] if args.key?(:stdout) @upload = args[:upload] if args.key?(:upload) @upload_start_time = args[:upload_start_time] if args.key?(:upload_start_time) end end # CommandEvents contains counters for the number of warnings and errors that # occurred during the execution of a command. class GoogleDevtoolsRemotebuildbotCommandEvents include Google::Apis::Core::Hashable # Indicates whether we are using a cached Docker image (true) or had to pull the # Docker image (false) for this command. # Corresponds to the JSON property `dockerCacheHit` # @return [Boolean] attr_accessor :docker_cache_hit alias_method :docker_cache_hit?, :docker_cache_hit # The input cache miss ratio. # Corresponds to the JSON property `inputCacheMiss` # @return [Float] attr_accessor :input_cache_miss # The number of errors reported. # Corresponds to the JSON property `numErrors` # @return [Fixnum] attr_accessor :num_errors # The number of warnings reported. # Corresponds to the JSON property `numWarnings` # @return [Fixnum] attr_accessor :num_warnings def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @docker_cache_hit = args[:docker_cache_hit] if args.key?(:docker_cache_hit) @input_cache_miss = args[:input_cache_miss] if args.key?(:input_cache_miss) @num_errors = args[:num_errors] if args.key?(:num_errors) @num_warnings = args[:num_warnings] if args.key?(:num_warnings) end end # The internal status of the command result. class GoogleDevtoolsRemotebuildbotCommandStatus include Google::Apis::Core::Hashable # The status code. # Corresponds to the JSON property `code` # @return [String] attr_accessor :code # The error message. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @code = args[:code] if args.key?(:code) @message = args[:message] if args.key?(:message) end end # ResourceUsage is the system resource usage of the host machine. class GoogleDevtoolsRemotebuildbotResourceUsage include Google::Apis::Core::Hashable # # Corresponds to the JSON property `cpuUsedPercent` # @return [Float] attr_accessor :cpu_used_percent # # Corresponds to the JSON property `diskUsage` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildbotResourceUsageStat] attr_accessor :disk_usage # # Corresponds to the JSON property `memoryUsage` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildbotResourceUsageStat] attr_accessor :memory_usage def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cpu_used_percent = args[:cpu_used_percent] if args.key?(:cpu_used_percent) @disk_usage = args[:disk_usage] if args.key?(:disk_usage) @memory_usage = args[:memory_usage] if args.key?(:memory_usage) end end # class GoogleDevtoolsRemotebuildbotResourceUsageStat include Google::Apis::Core::Hashable # # Corresponds to the JSON property `total` # @return [Fixnum] attr_accessor :total # # Corresponds to the JSON property `used` # @return [Fixnum] attr_accessor :used def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @total = args[:total] if args.key?(:total) @used = args[:used] if args.key?(:used) end end # AcceleratorConfig defines the accelerator cards to attach to the VM. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaAcceleratorConfig include Google::Apis::Core::Hashable # The number of guest accelerator cards exposed to each VM. # Corresponds to the JSON property `acceleratorCount` # @return [Fixnum] attr_accessor :accelerator_count # The type of accelerator to attach to each VM, e.g. "nvidia-tesla-k80" for # nVidia Tesla K80. # Corresponds to the JSON property `acceleratorType` # @return [String] attr_accessor :accelerator_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @accelerator_count = args[:accelerator_count] if args.key?(:accelerator_count) @accelerator_type = args[:accelerator_type] if args.key?(:accelerator_type) end end # Autoscale defines the autoscaling policy of a worker pool. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaAutoscale include Google::Apis::Core::Hashable # The maximal number of workers. Must be equal to or greater than min_size. # Corresponds to the JSON property `maxSize` # @return [Fixnum] attr_accessor :max_size # The minimal number of workers. Must be greater than 0. # Corresponds to the JSON property `minSize` # @return [Fixnum] attr_accessor :min_size def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @max_size = args[:max_size] if args.key?(:max_size) @min_size = args[:min_size] if args.key?(:min_size) end end # The request used for `CreateInstance`. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateInstanceRequest include Google::Apis::Core::Hashable # Instance conceptually encapsulates all Remote Build Execution resources for # remote builds. An instance consists of storage and compute resources (for # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for # running remote builds. All Remote Build Execution API calls are scoped to an # instance. # Corresponds to the JSON property `instance` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance] attr_accessor :instance # ID of the created instance. A valid `instance_id` must: be 6-50 characters # long, contain only lowercase letters, digits, hyphens and underscores, start # with a lowercase letter, and end with a lowercase letter or a digit. # Corresponds to the JSON property `instanceId` # @return [String] attr_accessor :instance_id # Resource name of the project containing the instance. Format: `projects/[ # PROJECT_ID]`. # Corresponds to the JSON property `parent` # @return [String] attr_accessor :parent def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance = args[:instance] if args.key?(:instance) @instance_id = args[:instance_id] if args.key?(:instance_id) @parent = args[:parent] if args.key?(:parent) end end # The request used for `CreateWorkerPool`. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateWorkerPoolRequest include Google::Apis::Core::Hashable # Resource name of the instance in which to create the new worker pool. Format: ` # projects/[PROJECT_ID]/instances/[INSTANCE_ID]`. # Corresponds to the JSON property `parent` # @return [String] attr_accessor :parent # ID of the created worker pool. A valid pool ID must: be 6-50 characters long, # contain only lowercase letters, digits, hyphens and underscores, start with a # lowercase letter, and end with a lowercase letter or a digit. # Corresponds to the JSON property `poolId` # @return [String] attr_accessor :pool_id # A worker pool resource in the Remote Build Execution API. # Corresponds to the JSON property `workerPool` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool] attr_accessor :worker_pool def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @parent = args[:parent] if args.key?(:parent) @pool_id = args[:pool_id] if args.key?(:pool_id) @worker_pool = args[:worker_pool] if args.key?(:worker_pool) end end # The request used for `DeleteInstance`. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteInstanceRequest include Google::Apis::Core::Hashable # Name of the instance to delete. Format: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) end end # The request used for DeleteWorkerPool. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteWorkerPoolRequest include Google::Apis::Core::Hashable # Name of the worker pool to delete. Format: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]/workerpools/[POOL_ID]`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) end end # FeaturePolicy defines features allowed to be used on RBE instances, as well as # instance-wide behavior changes that take effect without opt-in or opt-out at # usage time. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy include Google::Apis::Core::Hashable # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `containerImageSources` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :container_image_sources # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerAddCapabilities` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_add_capabilities # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerChrootPath` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_chroot_path # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerNetwork` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_network # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerPrivileged` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_privileged # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerRunAsRoot` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_run_as_root # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerRuntime` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_runtime # Defines whether a feature can be used or what values are accepted. # Corresponds to the JSON property `dockerSiblingContainers` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature] attr_accessor :docker_sibling_containers # linux_isolation allows overriding the docker runtime used for containers # started on Linux. # Corresponds to the JSON property `linuxIsolation` # @return [String] attr_accessor :linux_isolation def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @container_image_sources = args[:container_image_sources] if args.key?(:container_image_sources) @docker_add_capabilities = args[:docker_add_capabilities] if args.key?(:docker_add_capabilities) @docker_chroot_path = args[:docker_chroot_path] if args.key?(:docker_chroot_path) @docker_network = args[:docker_network] if args.key?(:docker_network) @docker_privileged = args[:docker_privileged] if args.key?(:docker_privileged) @docker_run_as_root = args[:docker_run_as_root] if args.key?(:docker_run_as_root) @docker_runtime = args[:docker_runtime] if args.key?(:docker_runtime) @docker_sibling_containers = args[:docker_sibling_containers] if args.key?(:docker_sibling_containers) @linux_isolation = args[:linux_isolation] if args.key?(:linux_isolation) end end # Defines whether a feature can be used or what values are accepted. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature include Google::Apis::Core::Hashable # A list of acceptable values. Only effective when the policy is `RESTRICTED`. # Corresponds to the JSON property `allowedValues` # @return [Array] attr_accessor :allowed_values # The policy of the feature. # Corresponds to the JSON property `policy` # @return [String] attr_accessor :policy def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @allowed_values = args[:allowed_values] if args.key?(:allowed_values) @policy = args[:policy] if args.key?(:policy) end end # The request used for `GetInstance`. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetInstanceRequest include Google::Apis::Core::Hashable # Name of the instance to retrieve. Format: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) end end # The request used for GetWorkerPool. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetWorkerPoolRequest include Google::Apis::Core::Hashable # Name of the worker pool to retrieve. Format: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]/workerpools/[POOL_ID]`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) end end # Instance conceptually encapsulates all Remote Build Execution resources for # remote builds. An instance consists of storage and compute resources (for # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for # running remote builds. All Remote Build Execution API calls are scoped to an # instance. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance include Google::Apis::Core::Hashable # FeaturePolicy defines features allowed to be used on RBE instances, as well as # instance-wide behavior changes that take effect without opt-in or opt-out at # usage time. # Corresponds to the JSON property `featurePolicy` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy] attr_accessor :feature_policy # The location is a GCP region. Currently only `us-central1` is supported. # Corresponds to the JSON property `location` # @return [String] attr_accessor :location # Output only. Whether stack driver logging is enabled for the instance. # Corresponds to the JSON property `loggingEnabled` # @return [Boolean] attr_accessor :logging_enabled alias_method :logging_enabled?, :logging_enabled # Output only. Instance resource name formatted as: `projects/[PROJECT_ID]/ # instances/[INSTANCE_ID]`. Name should not be populated when creating an # instance since it is provided in the `instance_id` field. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Output only. State of the instance. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @feature_policy = args[:feature_policy] if args.key?(:feature_policy) @location = args[:location] if args.key?(:location) @logging_enabled = args[:logging_enabled] if args.key?(:logging_enabled) @name = args[:name] if args.key?(:name) @state = args[:state] if args.key?(:state) end end # class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListInstancesRequest include Google::Apis::Core::Hashable # Resource name of the project. Format: `projects/[PROJECT_ID]`. # Corresponds to the JSON property `parent` # @return [String] attr_accessor :parent def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @parent = args[:parent] if args.key?(:parent) end end # class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListInstancesResponse include Google::Apis::Core::Hashable # The list of instances in a given project. # Corresponds to the JSON property `instances` # @return [Array] attr_accessor :instances def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instances = args[:instances] if args.key?(:instances) end end # class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListWorkerPoolsRequest include Google::Apis::Core::Hashable # Optional. A filter expression that filters resources listed in the response. # The expression must specify the field name, a comparison operator, and the # value that you want to use for filtering. The value must be a string, a number, # or a boolean. String values are case-insensitive. The comparison operator # must be either `:`, `=`, `!=`, `>`, `>=`, `<=` or `<`. The `:` operator can be # used with string fields to match substrings. For non-string fields it is # equivalent to the `=` operator. The `:*` comparison can be used to test # whether a key has been defined. You can also filter on nested fields. To # filter on multiple expressions, you can separate expression using `AND` and ` # OR` operators, using parentheses to specify precedence. If neither operator is # specified, `AND` is assumed. Examples: Include only pools with more than 100 # reserved workers: `(worker_count > 100) (worker_config.reserved = true)` # Include only pools with a certain label or machines of the n1-standard family: # `worker_config.labels.key1 : * OR worker_config.machine_type: n1-standard` # Corresponds to the JSON property `filter` # @return [String] attr_accessor :filter # Resource name of the instance. Format: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]`. # Corresponds to the JSON property `parent` # @return [String] attr_accessor :parent def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @filter = args[:filter] if args.key?(:filter) @parent = args[:parent] if args.key?(:parent) end end # class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListWorkerPoolsResponse include Google::Apis::Core::Hashable # The list of worker pools in a given instance. # Corresponds to the JSON property `workerPools` # @return [Array] attr_accessor :worker_pools def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @worker_pools = args[:worker_pools] if args.key?(:worker_pools) end end # SoleTenancyConfig specifies information required to host a pool on STNs. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig include Google::Apis::Core::Hashable # The sole-tenant node type to host the pool's workers on. # Corresponds to the JSON property `nodeType` # @return [String] attr_accessor :node_type # Zone in which STNs are reserved. # Corresponds to the JSON property `nodesZone` # @return [String] attr_accessor :nodes_zone def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @node_type = args[:node_type] if args.key?(:node_type) @nodes_zone = args[:nodes_zone] if args.key?(:nodes_zone) end end # The request used for `UpdateInstance`. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateInstanceRequest include Google::Apis::Core::Hashable # Instance conceptually encapsulates all Remote Build Execution resources for # remote builds. An instance consists of storage and compute resources (for # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for # running remote builds. All Remote Build Execution API calls are scoped to an # instance. # Corresponds to the JSON property `instance` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance] attr_accessor :instance # Deprecated, use instance.logging_enabled instead. Whether to enable # Stackdriver logging for this instance. # Corresponds to the JSON property `loggingEnabled` # @return [Boolean] attr_accessor :logging_enabled alias_method :logging_enabled?, :logging_enabled # Deprecated, use instance.Name instead. Name of the instance to update. Format: # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The update mask applies to instance. For the `FieldMask` definition, see https: # //developers.google.com/protocol-buffers/docs/reference/google.protobuf# # fieldmask If an empty update_mask is provided, only the non-default valued # field in the worker pool field will be updated. Note that in order to update a # field to the default value (zero, false, empty string) an explicit update_mask # must be provided. # Corresponds to the JSON property `updateMask` # @return [String] attr_accessor :update_mask def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance = args[:instance] if args.key?(:instance) @logging_enabled = args[:logging_enabled] if args.key?(:logging_enabled) @name = args[:name] if args.key?(:name) @update_mask = args[:update_mask] if args.key?(:update_mask) end end # The request used for UpdateWorkerPool. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateWorkerPoolRequest include Google::Apis::Core::Hashable # The update mask applies to worker_pool. For the `FieldMask` definition, see # https://developers.google.com/protocol-buffers/docs/reference/google.protobuf# # fieldmask If an empty update_mask is provided, only the non-default valued # field in the worker pool field will be updated. Note that in order to update a # field to the default value (zero, false, empty string) an explicit update_mask # must be provided. # Corresponds to the JSON property `updateMask` # @return [String] attr_accessor :update_mask # A worker pool resource in the Remote Build Execution API. # Corresponds to the JSON property `workerPool` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool] attr_accessor :worker_pool def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @update_mask = args[:update_mask] if args.key?(:update_mask) @worker_pool = args[:worker_pool] if args.key?(:worker_pool) end end # Defines the configuration to be used for creating workers in the worker pool. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig include Google::Apis::Core::Hashable # AcceleratorConfig defines the accelerator cards to attach to the VM. # Corresponds to the JSON property `accelerator` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaAcceleratorConfig] attr_accessor :accelerator # Required. Size of the disk attached to the worker, in GB. See https://cloud. # google.com/compute/docs/disks/ # Corresponds to the JSON property `diskSizeGb` # @return [Fixnum] attr_accessor :disk_size_gb # Required. Disk Type to use for the worker. See [Storage options](https://cloud. # google.com/compute/docs/disks/#introduction). Currently only `pd-standard` and # `pd-ssd` are supported. # Corresponds to the JSON property `diskType` # @return [String] attr_accessor :disk_type # Labels associated with the workers. Label keys and values can be no longer # than 63 characters, can only contain lowercase letters, numeric characters, # underscores and dashes. International letters are permitted. Label keys must # start with a letter. Label values are optional. There can not be more than 64 # labels per resource. # Corresponds to the JSON property `labels` # @return [Hash] attr_accessor :labels # Required. Machine type of the worker, such as `n1-standard-2`. See https:// # cloud.google.com/compute/docs/machine-types for a list of supported machine # types. Note that `f1-micro` and `g1-small` are not yet supported. # Corresponds to the JSON property `machineType` # @return [String] attr_accessor :machine_type # The maximum number of actions a worker can execute concurrently. # Corresponds to the JSON property `maxConcurrentActions` # @return [Fixnum] attr_accessor :max_concurrent_actions # Minimum CPU platform to use when creating the worker. See [CPU Platforms]( # https://cloud.google.com/compute/docs/cpu-platforms). # Corresponds to the JSON property `minCpuPlatform` # @return [String] attr_accessor :min_cpu_platform # Determines the type of network access granted to workers. Possible values: - " # public": Workers can connect to the public internet. - "private": Workers can # only connect to Google APIs and services. - "restricted-private": Workers can # only connect to Google APIs that are reachable through `restricted.googleapis. # com` (`199.36.153.4/30`). # Corresponds to the JSON property `networkAccess` # @return [String] attr_accessor :network_access # Determines whether the worker is reserved (equivalent to a Compute Engine on- # demand VM and therefore won't be preempted). See [Preemptible VMs](https:// # cloud.google.com/preemptible-vms/) for more details. # Corresponds to the JSON property `reserved` # @return [Boolean] attr_accessor :reserved alias_method :reserved?, :reserved # SoleTenancyConfig specifies information required to host a pool on STNs. # Corresponds to the JSON property `soleTenancy` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig] attr_accessor :sole_tenancy # The name of the image used by each VM. # Corresponds to the JSON property `vmImage` # @return [String] attr_accessor :vm_image def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @accelerator = args[:accelerator] if args.key?(:accelerator) @disk_size_gb = args[:disk_size_gb] if args.key?(:disk_size_gb) @disk_type = args[:disk_type] if args.key?(:disk_type) @labels = args[:labels] if args.key?(:labels) @machine_type = args[:machine_type] if args.key?(:machine_type) @max_concurrent_actions = args[:max_concurrent_actions] if args.key?(:max_concurrent_actions) @min_cpu_platform = args[:min_cpu_platform] if args.key?(:min_cpu_platform) @network_access = args[:network_access] if args.key?(:network_access) @reserved = args[:reserved] if args.key?(:reserved) @sole_tenancy = args[:sole_tenancy] if args.key?(:sole_tenancy) @vm_image = args[:vm_image] if args.key?(:vm_image) end end # A worker pool resource in the Remote Build Execution API. class GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool include Google::Apis::Core::Hashable # Autoscale defines the autoscaling policy of a worker pool. # Corresponds to the JSON property `autoscale` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaAutoscale] attr_accessor :autoscale # Channel specifies the release channel of the pool. # Corresponds to the JSON property `channel` # @return [String] attr_accessor :channel # WorkerPool resource name formatted as: `projects/[PROJECT_ID]/instances/[ # INSTANCE_ID]/workerpools/[POOL_ID]`. name should not be populated when # creating a worker pool since it is provided in the `poolId` field. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Output only. State of the worker pool. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state # Defines the configuration to be used for creating workers in the worker pool. # Corresponds to the JSON property `workerConfig` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig] attr_accessor :worker_config # The desired number of workers in the worker pool. Must be a value between 0 # and 15000. # Corresponds to the JSON property `workerCount` # @return [Fixnum] attr_accessor :worker_count def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @autoscale = args[:autoscale] if args.key?(:autoscale) @channel = args[:channel] if args.key?(:channel) @name = args[:name] if args.key?(:name) @state = args[:state] if args.key?(:state) @worker_config = args[:worker_config] if args.key?(:worker_config) @worker_count = args[:worker_count] if args.key?(:worker_count) end end # AdminTemp is a prelimiary set of administration tasks. It's called "Temp" # because we do not yet know the best way to represent admin tasks; it's # possible that this will be entirely replaced in later versions of this API. If # this message proves to be sufficient, it will be renamed in the alpha or beta # release of this API. This message (suitably marshalled into a protobuf.Any) # can be used as the inline_assignment field in a lease; the lease assignment # field should simply be `"admin"` in these cases. This message is heavily based # on Swarming administration tasks from the LUCI project (http://github.com/luci/ # luci-py/appengine/swarming). class GoogleDevtoolsRemoteworkersV1test2AdminTemp include Google::Apis::Core::Hashable # The argument to the admin action; see `Command` for semantics. # Corresponds to the JSON property `arg` # @return [String] attr_accessor :arg # The admin action; see `Command` for legal values. # Corresponds to the JSON property `command` # @return [String] attr_accessor :command def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @arg = args[:arg] if args.key?(:arg) @command = args[:command] if args.key?(:command) end end # Describes a blob of binary content with its digest. class GoogleDevtoolsRemoteworkersV1test2Blob include Google::Apis::Core::Hashable # The contents of the blob. # Corresponds to the JSON property `contents` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :contents # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest] attr_accessor :digest def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @contents = args[:contents] if args.key?(:contents) @digest = args[:digest] if args.key?(:digest) end end # DEPRECATED - use CommandResult instead. Describes the actual outputs from the # task. class GoogleDevtoolsRemoteworkersV1test2CommandOutputs include Google::Apis::Core::Hashable # exit_code is only fully reliable if the status' code is OK. If the task # exceeded its deadline or was cancelled, the process may still produce an exit # code as it is cancelled, and this will be populated, but a successful (zero) # is unlikely to be correct unless the status code is OK. # Corresponds to the JSON property `exitCode` # @return [Fixnum] attr_accessor :exit_code # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. # Corresponds to the JSON property `outputs` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest] attr_accessor :outputs def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @exit_code = args[:exit_code] if args.key?(:exit_code) @outputs = args[:outputs] if args.key?(:outputs) end end # DEPRECATED - use CommandResult instead. Can be used as part of CompleteRequest. # metadata, or are part of a more sophisticated message. class GoogleDevtoolsRemoteworkersV1test2CommandOverhead include Google::Apis::Core::Hashable # The elapsed time between calling Accept and Complete. The server will also # have its own idea of what this should be, but this excludes the overhead of # the RPCs and the bot response time. # Corresponds to the JSON property `duration` # @return [String] attr_accessor :duration # The amount of time *not* spent executing the command (ie uploading/downloading # files). # Corresponds to the JSON property `overhead` # @return [String] attr_accessor :overhead def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @duration = args[:duration] if args.key?(:duration) @overhead = args[:overhead] if args.key?(:overhead) end end # All information about the execution of a command, suitable for providing as # the Bots interface's `Lease.result` field. class GoogleDevtoolsRemoteworkersV1test2CommandResult include Google::Apis::Core::Hashable # The elapsed time between calling Accept and Complete. The server will also # have its own idea of what this should be, but this excludes the overhead of # the RPCs and the bot response time. # Corresponds to the JSON property `duration` # @return [String] attr_accessor :duration # The exit code of the process. An exit code of "0" should only be trusted if ` # status` has a code of OK (otherwise it may simply be unset). # Corresponds to the JSON property `exitCode` # @return [Fixnum] attr_accessor :exit_code # Implementation-dependent metadata about the task. Both servers and bots may # define messages which can be encoded here; bots are free to provide metadata # in multiple formats, and servers are free to choose one or more of the values # to process and ignore others. In particular, it is *not* considered an error # for the bot to provide the server with a field that it doesn't know about. # Corresponds to the JSON property `metadata` # @return [Array>] attr_accessor :metadata # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. # Corresponds to the JSON property `outputs` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest] attr_accessor :outputs # The amount of time *not* spent executing the command (ie uploading/downloading # files). # Corresponds to the JSON property `overhead` # @return [String] attr_accessor :overhead # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @duration = args[:duration] if args.key?(:duration) @exit_code = args[:exit_code] if args.key?(:exit_code) @metadata = args[:metadata] if args.key?(:metadata) @outputs = args[:outputs] if args.key?(:outputs) @overhead = args[:overhead] if args.key?(:overhead) @status = args[:status] if args.key?(:status) end end # Describes a shell-style task to execute, suitable for providing as the Bots # interface's `Lease.payload` field. class GoogleDevtoolsRemoteworkersV1test2CommandTask include Google::Apis::Core::Hashable # Describes the expected outputs of the command. # Corresponds to the JSON property `expectedOutputs` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskOutputs] attr_accessor :expected_outputs # Describes the inputs to a shell-style task. # Corresponds to the JSON property `inputs` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskInputs] attr_accessor :inputs # Describes the timeouts associated with this task. # Corresponds to the JSON property `timeouts` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskTimeouts] attr_accessor :timeouts def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @expected_outputs = args[:expected_outputs] if args.key?(:expected_outputs) @inputs = args[:inputs] if args.key?(:inputs) @timeouts = args[:timeouts] if args.key?(:timeouts) end end # Describes the inputs to a shell-style task. class GoogleDevtoolsRemoteworkersV1test2CommandTaskInputs include Google::Apis::Core::Hashable # The command itself to run (e.g., argv). This field should be passed directly # to the underlying operating system, and so it must be sensible to that # operating system. For example, on Windows, the first argument might be "C:\ # Windows\System32\ping.exe" - that is, using drive letters and backslashes. A # command for a *nix system, on the other hand, would use forward slashes. All # other fields in the RWAPI must consistently use forward slashes, since those # fields may be interpretted by both the service and the bot. # Corresponds to the JSON property `arguments` # @return [Array] attr_accessor :arguments # All environment variables required by the task. # Corresponds to the JSON property `environmentVariables` # @return [Array] attr_accessor :environment_variables # The input filesystem to be set up prior to the task beginning. The contents # should be a repeated set of FileMetadata messages though other formats are # allowed if better for the implementation (eg, a LUCI-style .isolated file). # This field is repeated since implementations might want to cache the metadata, # in which case it may be useful to break up portions of the filesystem that # change frequently (eg, specific input files) from those that don't (eg, # standard header files). # Corresponds to the JSON property `files` # @return [Array] attr_accessor :files # Inline contents for blobs expected to be needed by the bot to execute the task. # For example, contents of entries in `files` or blobs that are indirectly # referenced by an entry there. The bot should check against this list before # downloading required task inputs to reduce the number of communications # between itself and the remote CAS server. # Corresponds to the JSON property `inlineBlobs` # @return [Array] attr_accessor :inline_blobs # Directory from which a command is executed. It is a relative directory with # respect to the bot's working directory (i.e., "./"). If it is non-empty, then # it must exist under "./". Otherwise, "./" will be used. # Corresponds to the JSON property `workingDirectory` # @return [String] attr_accessor :working_directory def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @arguments = args[:arguments] if args.key?(:arguments) @environment_variables = args[:environment_variables] if args.key?(:environment_variables) @files = args[:files] if args.key?(:files) @inline_blobs = args[:inline_blobs] if args.key?(:inline_blobs) @working_directory = args[:working_directory] if args.key?(:working_directory) end end # An environment variable required by this task. class GoogleDevtoolsRemoteworkersV1test2CommandTaskInputsEnvironmentVariable include Google::Apis::Core::Hashable # The envvar name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The envvar value. # Corresponds to the JSON property `value` # @return [String] attr_accessor :value def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @value = args[:value] if args.key?(:value) end end # Describes the expected outputs of the command. class GoogleDevtoolsRemoteworkersV1test2CommandTaskOutputs include Google::Apis::Core::Hashable # A list of expected directories, relative to the execution root. All paths MUST # be delimited by forward slashes. # Corresponds to the JSON property `directories` # @return [Array] attr_accessor :directories # A list of expected files, relative to the execution root. All paths MUST be # delimited by forward slashes. # Corresponds to the JSON property `files` # @return [Array] attr_accessor :files # The destination to which any stderr should be sent. The method by which the # bot should send the stream contents to that destination is not defined in this # API. As examples, the destination could be a file referenced in the `files` # field in this message, or it could be a URI that must be written via the # ByteStream API. # Corresponds to the JSON property `stderrDestination` # @return [String] attr_accessor :stderr_destination # The destination to which any stdout should be sent. The method by which the # bot should send the stream contents to that destination is not defined in this # API. As examples, the destination could be a file referenced in the `files` # field in this message, or it could be a URI that must be written via the # ByteStream API. # Corresponds to the JSON property `stdoutDestination` # @return [String] attr_accessor :stdout_destination def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @directories = args[:directories] if args.key?(:directories) @files = args[:files] if args.key?(:files) @stderr_destination = args[:stderr_destination] if args.key?(:stderr_destination) @stdout_destination = args[:stdout_destination] if args.key?(:stdout_destination) end end # Describes the timeouts associated with this task. class GoogleDevtoolsRemoteworkersV1test2CommandTaskTimeouts include Google::Apis::Core::Hashable # This specifies the maximum time that the task can run, excluding the time # required to download inputs or upload outputs. That is, the worker will # terminate the task if it runs longer than this. # Corresponds to the JSON property `execution` # @return [String] attr_accessor :execution # This specifies the maximum amount of time the task can be idle - that is, go # without generating some output in either stdout or stderr. If the process is # silent for more than the specified time, the worker will terminate the task. # Corresponds to the JSON property `idle` # @return [String] attr_accessor :idle # If the execution or IO timeouts are exceeded, the worker will try to # gracefully terminate the task and return any existing logs. However, tasks may # be hard-frozen in which case this process will fail. This timeout specifies # how long to wait for a terminated task to shut down gracefully (e.g. via # SIGTERM) before we bring down the hammer (e.g. SIGKILL on *nix, # CTRL_BREAK_EVENT on Windows). # Corresponds to the JSON property `shutdown` # @return [String] attr_accessor :shutdown def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @execution = args[:execution] if args.key?(:execution) @idle = args[:idle] if args.key?(:idle) @shutdown = args[:shutdown] if args.key?(:shutdown) end end # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. class GoogleDevtoolsRemoteworkersV1test2Digest include Google::Apis::Core::Hashable # A string-encoded hash (eg "1a2b3c", not the byte array [0x1a, 0x2b, 0x3c]) # using an implementation-defined hash algorithm (eg SHA-256). # Corresponds to the JSON property `hash` # @return [String] attr_accessor :hash_prop # The size of the contents. While this is not strictly required as part of an # identifier (after all, any given hash will have exactly one canonical size), # it's useful in almost all cases when one might want to send or retrieve blobs # of content and is included here for this reason. # Corresponds to the JSON property `sizeBytes` # @return [Fixnum] attr_accessor :size_bytes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @hash_prop = args[:hash_prop] if args.key?(:hash_prop) @size_bytes = args[:size_bytes] if args.key?(:size_bytes) end end # The contents of a directory. Similar to the equivalent message in the Remote # Execution API. class GoogleDevtoolsRemoteworkersV1test2Directory include Google::Apis::Core::Hashable # Any subdirectories # Corresponds to the JSON property `directories` # @return [Array] attr_accessor :directories # The files in this directory # Corresponds to the JSON property `files` # @return [Array] attr_accessor :files def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @directories = args[:directories] if args.key?(:directories) @files = args[:files] if args.key?(:files) end end # The metadata for a directory. Similar to the equivalent message in the Remote # Execution API. class GoogleDevtoolsRemoteworkersV1test2DirectoryMetadata include Google::Apis::Core::Hashable # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest] attr_accessor :digest # The path of the directory, as in FileMetadata.path. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @digest = args[:digest] if args.key?(:digest) @path = args[:path] if args.key?(:path) end end # The metadata for a file. Similar to the equivalent message in the Remote # Execution API. class GoogleDevtoolsRemoteworkersV1test2FileMetadata include Google::Apis::Core::Hashable # If the file is small enough, its contents may also or alternatively be listed # here. # Corresponds to the JSON property `contents` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :contents # The CommandTask and CommandResult messages assume the existence of a service # that can serve blobs of content, identified by a hash and size known as a " # digest." The method by which these blobs may be retrieved is not specified # here, but a model implementation is in the Remote Execution API's " # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest # will virtually always refer to the contents of a file or a directory. The # latter is represented by the byte-encoded Directory message. # Corresponds to the JSON property `digest` # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest] attr_accessor :digest # Properties of the file # Corresponds to the JSON property `isExecutable` # @return [Boolean] attr_accessor :is_executable alias_method :is_executable?, :is_executable # The path of this file. If this message is part of the CommandOutputs.outputs # fields, the path is relative to the execution root and must correspond to an # entry in CommandTask.outputs.files. If this message is part of a Directory # message, then the path is relative to the root of that directory. All paths # MUST be delimited by forward slashes. # Corresponds to the JSON property `path` # @return [String] attr_accessor :path def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @contents = args[:contents] if args.key?(:contents) @digest = args[:digest] if args.key?(:digest) @is_executable = args[:is_executable] if args.key?(:is_executable) @path = args[:path] if args.key?(:path) end end # This resource represents a long-running operation that is the result of a # network API call. class GoogleLongrunningOperation include Google::Apis::Core::Hashable # If the value is `false`, it means the operation is still in progress. If `true` # , the operation is completed, and either `error` or `response` is available. # Corresponds to the JSON property `done` # @return [Boolean] attr_accessor :done alias_method :done?, :done # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `error` # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus] attr_accessor :error # Service-specific metadata associated with the operation. It typically contains # progress information and common metadata such as create time. Some services # might not provide such metadata. Any method that returns a long-running # operation should document the metadata type, if any. # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # The server-assigned name, which is only unique within the same service that # originally returns it. If you use the default HTTP mapping, the `name` should # be a resource name ending with `operations/`unique_id``. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The normal response of the operation in case of success. If the original # method returns no data on success, such as `Delete`, the response is `google. # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, # the response should be the resource. For other methods, the response should # have the type `XxxResponse`, where `Xxx` is the original method name. For # example, if the original method name is `TakeSnapshot()`, the inferred # response type is `TakeSnapshotResponse`. # Corresponds to the JSON property `response` # @return [Hash] attr_accessor :response def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @done = args[:done] if args.key?(:done) @error = args[:error] if args.key?(:error) @metadata = args[:metadata] if args.key?(:metadata) @name = args[:name] if args.key?(:name) @response = args[:response] if args.key?(:response) end end # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). class GoogleRpcStatus include Google::Apis::Core::Hashable # The status code, which should be an enum value of google.rpc.Code. # Corresponds to the JSON property `code` # @return [Fixnum] attr_accessor :code # A list of messages that carry the error details. There is a common set of # message types for APIs to use. # Corresponds to the JSON property `details` # @return [Array>] attr_accessor :details # A developer-facing error message, which should be in English. Any user-facing # error message should be localized and sent in the google.rpc.Status.details # field, or localized by the client. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @code = args[:code] if args.key?(:code) @details = args[:details] if args.key?(:details) @message = args[:message] if args.key?(:message) end end end end end