# Copyright 2015 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

require 'date'
require 'google/apis/core/base_service'
require 'google/apis/core/json_representation'
require 'google/apis/core/hashable'
require 'google/apis/errors'

module Google
  module Apis
    module RemotebuildexecutionV2
      
      # An `Action` captures all the information about an execution which is required
      # to reproduce it. `Action`s are the core component of the [Execution] service.
      # A single `Action` represents a repeatable action that can be performed by the
      # execution service. `Action`s can be succinctly identified by the digest of
      # their wire format encoding and, once an `Action` has been executed, will be
      # cached in the action cache. Future requests can then use the cached result
      # rather than needing to run afresh. When a server completes execution of an
      # Action, it MAY choose to cache the result in the ActionCache unless `
      # do_not_cache` is `true`. Clients SHOULD expect the server to do so. By default,
      # future calls to Execute the same `Action` will also serve their results from
      # the cache. Clients must take care to understand the caching behaviour. Ideally,
      # all `Action`s will be reproducible so that serving a result from cache is
      # always desirable and correct.
      class BuildBazelRemoteExecutionV2Action
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `commandDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :command_digest
      
        # If true, then the `Action`'s result cannot be cached, and in-flight requests
        # for the same `Action` may not be merged.
        # Corresponds to the JSON property `doNotCache`
        # @return [Boolean]
        attr_accessor :do_not_cache
        alias_method :do_not_cache?, :do_not_cache
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `inputRootDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :input_root_digest
      
        # List of required supported NodeProperty keys. In order to ensure that
        # equivalent `Action`s always hash to the same value, the supported node
        # properties MUST be lexicographically sorted by name. Sorting of strings is
        # done by code point, equivalently, by the UTF-8 bytes. The interpretation of
        # these properties is server-dependent. If a property is not recognized by the
        # server, the server will return an `INVALID_ARGUMENT` error.
        # Corresponds to the JSON property `outputNodeProperties`
        # @return [Array<String>]
        attr_accessor :output_node_properties
      
        # A timeout after which the execution should be killed. If the timeout is absent,
        # then the client is specifying that the execution should continue as long as
        # the server will let it. The server SHOULD impose a timeout if the client does
        # not specify one, however, if the client does specify a timeout that is longer
        # than the server's maximum timeout, the server MUST reject the request. The
        # timeout is a part of the Action message, and therefore two `Actions` with
        # different timeouts are different, even if they are otherwise identical. This
        # is because, if they were not, running an `Action` with a lower timeout than is
        # required might result in a cache hit from an execution run with a longer
        # timeout, hiding the fact that the timeout is too short. By encoding it
        # directly in the `Action`, a lower timeout will result in a cache miss and the
        # execution timeout will fail immediately, rather than whenever the cache entry
        # gets evicted.
        # Corresponds to the JSON property `timeout`
        # @return [String]
        attr_accessor :timeout
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @command_digest = args[:command_digest] if args.key?(:command_digest)
          @do_not_cache = args[:do_not_cache] if args.key?(:do_not_cache)
          @input_root_digest = args[:input_root_digest] if args.key?(:input_root_digest)
          @output_node_properties = args[:output_node_properties] if args.key?(:output_node_properties)
          @timeout = args[:timeout] if args.key?(:timeout)
        end
      end
      
      # Describes the server/instance capabilities for updating the action cache.
      class BuildBazelRemoteExecutionV2ActionCacheUpdateCapabilities
        include Google::Apis::Core::Hashable
      
        # 
        # Corresponds to the JSON property `updateEnabled`
        # @return [Boolean]
        attr_accessor :update_enabled
        alias_method :update_enabled?, :update_enabled
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @update_enabled = args[:update_enabled] if args.key?(:update_enabled)
        end
      end
      
      # An ActionResult represents the result of an Action being run.
      class BuildBazelRemoteExecutionV2ActionResult
        include Google::Apis::Core::Hashable
      
        # ExecutedActionMetadata contains details about a completed execution.
        # Corresponds to the JSON property `executionMetadata`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutedActionMetadata]
        attr_accessor :execution_metadata
      
        # The exit code of the command.
        # Corresponds to the JSON property `exitCode`
        # @return [Fixnum]
        attr_accessor :exit_code
      
        # The output directories of the action. For each output directory requested in
        # the `output_directories` or `output_paths` field of the Action, if the
        # corresponding directory existed after the action completed, a single entry
        # will be present in the output list, which will contain the digest of a Tree
        # message containing the directory tree, and the path equal exactly to the
        # corresponding Action output_directories member. As an example, suppose the
        # Action had an output directory `a/b/dir` and the execution produced the
        # following contents in `a/b/dir`: a file named `bar` and a directory named `foo`
        # with an executable file named `baz`. Then, output_directory will contain (
        # hashes shortened for readability): ```json // OutputDirectory proto: ` path: "
        # a/b/dir" tree_digest: ` hash: "4a73bc9d03...", size: 55 ` ` // Tree proto with
        # hash "4a73bc9d03..." and size 55: ` root: ` files: [ ` name: "bar", digest: `
        # hash: "4a73bc9d03...", size: 65534 ` ` ], directories: [ ` name: "foo", digest:
        # ` hash: "4cf2eda940...", size: 43 ` ` ] ` children : ` // (Directory proto
        # with hash "4cf2eda940..." and size 43) files: [ ` name: "baz", digest: ` hash:
        # "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ` ``` If an output
        # of the same name as listed in `output_files` of the Command was found in `
        # output_directories`, but was not a directory, the server will return a
        # FAILED_PRECONDITION.
        # Corresponds to the JSON property `outputDirectories`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2OutputDirectory>]
        attr_accessor :output_directories
      
        # The output directories of the action that are symbolic links to other
        # directories. Those may be links to other output directories, or input
        # directories, or even absolute paths outside of the working directory, if the
        # server supports SymlinkAbsolutePathStrategy.ALLOWED. For each output directory
        # requested in the `output_directories` field of the Action, if the directory
        # existed after the action completed, a single entry will be present either in
        # this field, or in the `output_directories` field, if the directory was not a
        # symbolic link. If an output of the same name was found, but was a symbolic
        # link to a file instead of a directory, the server will return a
        # FAILED_PRECONDITION. If the action does not produce the requested output, then
        # that output will be omitted from the list. The server is free to arrange the
        # output list as desired; clients MUST NOT assume that the output list is sorted.
        # DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API
        # should still populate this field in addition to `output_symlinks`.
        # Corresponds to the JSON property `outputDirectorySymlinks`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2OutputSymlink>]
        attr_accessor :output_directory_symlinks
      
        # The output files of the action that are symbolic links to other files. Those
        # may be links to other output files, or input files, or even absolute paths
        # outside of the working directory, if the server supports
        # SymlinkAbsolutePathStrategy.ALLOWED. For each output file requested in the `
        # output_files` or `output_paths` field of the Action, if the corresponding file
        # existed after the action completed, a single entry will be present either in
        # this field, or in the `output_files` field, if the file was not a symbolic
        # link. If an output symbolic link of the same name as listed in `output_files`
        # of the Command was found, but its target type was not a regular file, the
        # server will return a FAILED_PRECONDITION. If the action does not produce the
        # requested output, then that output will be omitted from the list. The server
        # is free to arrange the output list as desired; clients MUST NOT assume that
        # the output list is sorted. DEPRECATED as of v2.1. Servers that wish to be
        # compatible with v2.0 API should still populate this field in addition to `
        # output_symlinks`.
        # Corresponds to the JSON property `outputFileSymlinks`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2OutputSymlink>]
        attr_accessor :output_file_symlinks
      
        # The output files of the action. For each output file requested in the `
        # output_files` or `output_paths` field of the Action, if the corresponding file
        # existed after the action completed, a single entry will be present either in
        # this field, or the `output_file_symlinks` field if the file was a symbolic
        # link to another file (`output_symlinks` field after v2.1). If an output listed
        # in `output_files` was found, but was a directory rather than a regular file,
        # the server will return a FAILED_PRECONDITION. If the action does not produce
        # the requested output, then that output will be omitted from the list. The
        # server is free to arrange the output list as desired; clients MUST NOT assume
        # that the output list is sorted.
        # Corresponds to the JSON property `outputFiles`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2OutputFile>]
        attr_accessor :output_files
      
        # New in v2.1: this field will only be populated if the command `output_paths`
        # field was used, and not the pre v2.1 `output_files` or `output_directories`
        # fields. The output paths of the action that are symbolic links to other paths.
        # Those may be links to other outputs, or inputs, or even absolute paths outside
        # of the working directory, if the server supports SymlinkAbsolutePathStrategy.
        # ALLOWED. A single entry for each output requested in `output_paths` field of
        # the Action, if the corresponding path existed after the action completed and
        # was a symbolic link. If the action does not produce a requested output, then
        # that output will be omitted from the list. The server is free to arrange the
        # output list as desired; clients MUST NOT assume that the output list is sorted.
        # Corresponds to the JSON property `outputSymlinks`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2OutputSymlink>]
        attr_accessor :output_symlinks
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `stderrDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :stderr_digest
      
        # The standard error buffer of the action. The server SHOULD NOT inline stderr
        # unless requested by the client in the GetActionResultRequest message. The
        # server MAY omit inlining, even if requested, and MUST do so if inlining would
        # cause the response to exceed message size limits.
        # Corresponds to the JSON property `stderrRaw`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :stderr_raw
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `stdoutDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :stdout_digest
      
        # The standard output buffer of the action. The server SHOULD NOT inline stdout
        # unless requested by the client in the GetActionResultRequest message. The
        # server MAY omit inlining, even if requested, and MUST do so if inlining would
        # cause the response to exceed message size limits.
        # Corresponds to the JSON property `stdoutRaw`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :stdout_raw
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @execution_metadata = args[:execution_metadata] if args.key?(:execution_metadata)
          @exit_code = args[:exit_code] if args.key?(:exit_code)
          @output_directories = args[:output_directories] if args.key?(:output_directories)
          @output_directory_symlinks = args[:output_directory_symlinks] if args.key?(:output_directory_symlinks)
          @output_file_symlinks = args[:output_file_symlinks] if args.key?(:output_file_symlinks)
          @output_files = args[:output_files] if args.key?(:output_files)
          @output_symlinks = args[:output_symlinks] if args.key?(:output_symlinks)
          @stderr_digest = args[:stderr_digest] if args.key?(:stderr_digest)
          @stderr_raw = args[:stderr_raw] if args.key?(:stderr_raw)
          @stdout_digest = args[:stdout_digest] if args.key?(:stdout_digest)
          @stdout_raw = args[:stdout_raw] if args.key?(:stdout_raw)
        end
      end
      
      # A request message for ContentAddressableStorage.BatchReadBlobs.
      class BuildBazelRemoteExecutionV2BatchReadBlobsRequest
        include Google::Apis::Core::Hashable
      
        # The individual blob digests.
        # Corresponds to the JSON property `digests`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest>]
        attr_accessor :digests
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digests = args[:digests] if args.key?(:digests)
        end
      end
      
      # A response message for ContentAddressableStorage.BatchReadBlobs.
      class BuildBazelRemoteExecutionV2BatchReadBlobsResponse
        include Google::Apis::Core::Hashable
      
        # The responses to the requests.
        # Corresponds to the JSON property `responses`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2BatchReadBlobsResponseResponse>]
        attr_accessor :responses
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @responses = args[:responses] if args.key?(:responses)
        end
      end
      
      # A response corresponding to a single blob that the client tried to download.
      class BuildBazelRemoteExecutionV2BatchReadBlobsResponseResponse
        include Google::Apis::Core::Hashable
      
        # The raw binary data.
        # Corresponds to the JSON property `data`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :data
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # The `Status` type defines a logical error model that is suitable for different
        # programming environments, including REST APIs and RPC APIs. It is used by [
        # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
        # data: error code, error message, and error details. You can find out more
        # about this error model and how to work with it in the [API Design Guide](https:
        # //cloud.google.com/apis/design/errors).
        # Corresponds to the JSON property `status`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus]
        attr_accessor :status
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @data = args[:data] if args.key?(:data)
          @digest = args[:digest] if args.key?(:digest)
          @status = args[:status] if args.key?(:status)
        end
      end
      
      # A request message for ContentAddressableStorage.BatchUpdateBlobs.
      class BuildBazelRemoteExecutionV2BatchUpdateBlobsRequest
        include Google::Apis::Core::Hashable
      
        # The individual upload requests.
        # Corresponds to the JSON property `requests`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest>]
        attr_accessor :requests
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @requests = args[:requests] if args.key?(:requests)
        end
      end
      
      # A request corresponding to a single blob that the client wants to upload.
      class BuildBazelRemoteExecutionV2BatchUpdateBlobsRequestRequest
        include Google::Apis::Core::Hashable
      
        # The raw binary data.
        # Corresponds to the JSON property `data`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :data
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @data = args[:data] if args.key?(:data)
          @digest = args[:digest] if args.key?(:digest)
        end
      end
      
      # A response message for ContentAddressableStorage.BatchUpdateBlobs.
      class BuildBazelRemoteExecutionV2BatchUpdateBlobsResponse
        include Google::Apis::Core::Hashable
      
        # The responses to the requests.
        # Corresponds to the JSON property `responses`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2BatchUpdateBlobsResponseResponse>]
        attr_accessor :responses
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @responses = args[:responses] if args.key?(:responses)
        end
      end
      
      # A response corresponding to a single blob that the client tried to upload.
      class BuildBazelRemoteExecutionV2BatchUpdateBlobsResponseResponse
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # The `Status` type defines a logical error model that is suitable for different
        # programming environments, including REST APIs and RPC APIs. It is used by [
        # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
        # data: error code, error message, and error details. You can find out more
        # about this error model and how to work with it in the [API Design Guide](https:
        # //cloud.google.com/apis/design/errors).
        # Corresponds to the JSON property `status`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus]
        attr_accessor :status
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest = args[:digest] if args.key?(:digest)
          @status = args[:status] if args.key?(:status)
        end
      end
      
      # Capabilities of the remote cache system.
      class BuildBazelRemoteExecutionV2CacheCapabilities
        include Google::Apis::Core::Hashable
      
        # Describes the server/instance capabilities for updating the action cache.
        # Corresponds to the JSON property `actionCacheUpdateCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ActionCacheUpdateCapabilities]
        attr_accessor :action_cache_update_capabilities
      
        # Allowed values for priority in ResultsCachePolicy Used for querying both cache
        # and execution valid priority ranges.
        # Corresponds to the JSON property `cachePriorityCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PriorityCapabilities]
        attr_accessor :cache_priority_capabilities
      
        # All the digest functions supported by the remote cache. Remote cache may
        # support multiple digest functions simultaneously.
        # Corresponds to the JSON property `digestFunction`
        # @return [Array<String>]
        attr_accessor :digest_function
      
        # Maximum total size of blobs to be uploaded/downloaded using batch methods. A
        # value of 0 means no limit is set, although in practice there will always be a
        # message size limitation of the protocol in use, e.g. GRPC.
        # Corresponds to the JSON property `maxBatchTotalSizeBytes`
        # @return [Fixnum]
        attr_accessor :max_batch_total_size_bytes
      
        # Whether absolute symlink targets are supported.
        # Corresponds to the JSON property `symlinkAbsolutePathStrategy`
        # @return [String]
        attr_accessor :symlink_absolute_path_strategy
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @action_cache_update_capabilities = args[:action_cache_update_capabilities] if args.key?(:action_cache_update_capabilities)
          @cache_priority_capabilities = args[:cache_priority_capabilities] if args.key?(:cache_priority_capabilities)
          @digest_function = args[:digest_function] if args.key?(:digest_function)
          @max_batch_total_size_bytes = args[:max_batch_total_size_bytes] if args.key?(:max_batch_total_size_bytes)
          @symlink_absolute_path_strategy = args[:symlink_absolute_path_strategy] if args.key?(:symlink_absolute_path_strategy)
        end
      end
      
      # A `Command` is the actual command executed by a worker running an Action and
      # specifications of its environment. Except as otherwise required, the
      # environment (such as which system libraries or binaries are available, and
      # what filesystems are mounted where) is defined by and specific to the
      # implementation of the remote execution API.
      class BuildBazelRemoteExecutionV2Command
        include Google::Apis::Core::Hashable
      
        # The arguments to the command. The first argument must be the path to the
        # executable, which must be either a relative path, in which case it is
        # evaluated with respect to the input root, or an absolute path.
        # Corresponds to the JSON property `arguments`
        # @return [Array<String>]
        attr_accessor :arguments
      
        # The environment variables to set when running the program. The worker may
        # provide its own default environment variables; these defaults can be
        # overridden using this field. Additional variables can also be specified. In
        # order to ensure that equivalent Commands always hash to the same value, the
        # environment variables MUST be lexicographically sorted by name. Sorting of
        # strings is done by code point, equivalently, by the UTF-8 bytes.
        # Corresponds to the JSON property `environmentVariables`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2CommandEnvironmentVariable>]
        attr_accessor :environment_variables
      
        # A list of the output directories that the client expects to retrieve from the
        # action. Only the listed directories will be returned (an entire directory
        # structure will be returned as a Tree message digest, see OutputDirectory), as
        # well as files listed in `output_files`. Other files or directories that may be
        # created during command execution are discarded. The paths are relative to the
        # working directory of the action execution. The paths are specified using a
        # single forward slash (`/`) as a path separator, even if the execution platform
        # natively uses a different separator. The path MUST NOT include a trailing
        # slash, nor a leading slash, being a relative path. The special value of empty
        # string is allowed, although not recommended, and can be used to capture the
        # entire working directory tree, including inputs. In order to ensure consistent
        # hashing of the same Action, the output paths MUST be sorted lexicographically
        # by code point (or, equivalently, by UTF-8 bytes). An output directory cannot
        # be duplicated or have the same path as any of the listed output files. An
        # output directory is allowed to be a parent of another output directory.
        # Directories leading up to the output directories (but not the output
        # directories themselves) are created by the worker prior to execution, even if
        # they are not explicitly part of the input root. DEPRECATED since 2.1: Use `
        # output_paths` instead.
        # Corresponds to the JSON property `outputDirectories`
        # @return [Array<String>]
        attr_accessor :output_directories
      
        # A list of the output files that the client expects to retrieve from the action.
        # Only the listed files, as well as directories listed in `output_directories`,
        # will be returned to the client as output. Other files or directories that may
        # be created during command execution are discarded. The paths are relative to
        # the working directory of the action execution. The paths are specified using a
        # single forward slash (`/`) as a path separator, even if the execution platform
        # natively uses a different separator. The path MUST NOT include a trailing
        # slash, nor a leading slash, being a relative path. In order to ensure
        # consistent hashing of the same Action, the output paths MUST be sorted
        # lexicographically by code point (or, equivalently, by UTF-8 bytes). An output
        # file cannot be duplicated, be a parent of another output file, or have the
        # same path as any of the listed output directories. Directories leading up to
        # the output files are created by the worker prior to execution, even if they
        # are not explicitly part of the input root. DEPRECATED since v2.1: Use `
        # output_paths` instead.
        # Corresponds to the JSON property `outputFiles`
        # @return [Array<String>]
        attr_accessor :output_files
      
        # A list of the output paths that the client expects to retrieve from the action.
        # Only the listed paths will be returned to the client as output. The type of
        # the output (file or directory) is not specified, and will be determined by the
        # server after action execution. If the resulting path is a file, it will be
        # returned in an OutputFile) typed field. If the path is a directory, the entire
        # directory structure will be returned as a Tree message digest, see
        # OutputDirectory) Other files or directories that may be created during command
        # execution are discarded. The paths are relative to the working directory of
        # the action execution. The paths are specified using a single forward slash (`/`
        # ) as a path separator, even if the execution platform natively uses a
        # different separator. The path MUST NOT include a trailing slash, nor a leading
        # slash, being a relative path. In order to ensure consistent hashing of the
        # same Action, the output paths MUST be deduplicated and sorted
        # lexicographically by code point (or, equivalently, by UTF-8 bytes).
        # Directories leading up to the output paths are created by the worker prior to
        # execution, even if they are not explicitly part of the input root. New in v2.1:
        # this field supersedes the DEPRECATED `output_files` and `output_directories`
        # fields. If `output_paths` is used, `output_files` and `output_directories`
        # will be ignored!
        # Corresponds to the JSON property `outputPaths`
        # @return [Array<String>]
        attr_accessor :output_paths
      
        # A `Platform` is a set of requirements, such as hardware, operating system, or
        # compiler toolchain, for an Action's execution environment. A `Platform` is
        # represented as a series of key-value pairs representing the properties that
        # are required of the platform.
        # Corresponds to the JSON property `platform`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Platform]
        attr_accessor :platform
      
        # The working directory, relative to the input root, for the command to run in.
        # It must be a directory which exists in the input tree. If it is left empty,
        # then the action is run in the input root.
        # Corresponds to the JSON property `workingDirectory`
        # @return [String]
        attr_accessor :working_directory
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @arguments = args[:arguments] if args.key?(:arguments)
          @environment_variables = args[:environment_variables] if args.key?(:environment_variables)
          @output_directories = args[:output_directories] if args.key?(:output_directories)
          @output_files = args[:output_files] if args.key?(:output_files)
          @output_paths = args[:output_paths] if args.key?(:output_paths)
          @platform = args[:platform] if args.key?(:platform)
          @working_directory = args[:working_directory] if args.key?(:working_directory)
        end
      end
      
      # An `EnvironmentVariable` is one variable to set in the running program's
      # environment.
      class BuildBazelRemoteExecutionV2CommandEnvironmentVariable
        include Google::Apis::Core::Hashable
      
        # The variable name.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The variable value.
        # Corresponds to the JSON property `value`
        # @return [String]
        attr_accessor :value
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
          @value = args[:value] if args.key?(:value)
        end
      end
      
      # A content digest. A digest for a given blob consists of the size of the blob
      # and its hash. The hash algorithm to use is defined by the server. The size is
      # considered to be an integral part of the digest and cannot be separated. That
      # is, even if the `hash` field is correctly specified but `size_bytes` is not,
      # the server MUST reject the request. The reason for including the size in the
      # digest is as follows: in a great many cases, the server needs to know the size
      # of the blob it is about to work with prior to starting an operation with it,
      # such as flattening Merkle tree structures or streaming it to a worker.
      # Technically, the server could implement a separate metadata store, but this
      # results in a significantly more complicated implementation as opposed to
      # having the client specify the size up-front (or storing the size along with
      # the digest in every message where digests are embedded). This does mean that
      # the API leaks some implementation details of (what we consider to be) a
      # reasonable server implementation, but we consider this to be a worthwhile
      # tradeoff. When a `Digest` is used to refer to a proto message, it always
      # refers to the message in binary encoded form. To ensure consistent hashing,
      # clients and servers MUST ensure that they serialize messages according to the
      # following rules, even if there are alternate valid encodings for the same
      # message: * Fields are serialized in tag order. * There are no unknown fields. *
      # There are no duplicate fields. * Fields are serialized according to the
      # default semantics for their type. Most protocol buffer implementations will
      # always follow these rules when serializing, but care should be taken to avoid
      # shortcuts. For instance, concatenating two messages to merge them may produce
      # duplicate fields.
      class BuildBazelRemoteExecutionV2Digest
        include Google::Apis::Core::Hashable
      
        # The hash. In the case of SHA-256, it will always be a lowercase hex string
        # exactly 64 characters long.
        # Corresponds to the JSON property `hash`
        # @return [String]
        attr_accessor :hash_prop
      
        # The size of the blob, in bytes.
        # Corresponds to the JSON property `sizeBytes`
        # @return [Fixnum]
        attr_accessor :size_bytes
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @hash_prop = args[:hash_prop] if args.key?(:hash_prop)
          @size_bytes = args[:size_bytes] if args.key?(:size_bytes)
        end
      end
      
      # A `Directory` represents a directory node in a file tree, containing zero or
      # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains
      # its name in the directory, either the digest of its content (either a file
      # blob or a `Directory` proto) or a symlink target, as well as possibly some
      # metadata about the file or directory. In order to ensure that two equivalent
      # directory trees hash to the same value, the following restrictions MUST be
      # obeyed when constructing a a `Directory`: * Every child in the directory must
      # have a path of exactly one segment. Multiple levels of directory hierarchy may
      # not be collapsed. * Each child in the directory must have a unique path
      # segment (file name). Note that while the API itself is case-sensitive, the
      # environment where the Action is executed may or may not be case-sensitive.
      # That is, it is legal to call the API with a Directory that has both "Foo" and "
      # foo" as children, but the Action may be rejected by the remote system upon
      # execution. * The files, directories and symlinks in the directory must each be
      # sorted in lexicographical order by path. The path strings must be sorted by
      # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files,
      # directories, and symlinks must be sorted in lexicographical order by property
      # name. A `Directory` that obeys the restrictions is said to be in canonical
      # form. As an example, the following could be used for a file named `bar` and a
      # directory named `foo` with an executable file named `baz` (hashes shortened
      # for readability): ```json // (Directory proto) ` files: [ ` name: "bar",
      # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": "
      # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: "
      # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto
      # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: `
      # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ```
      class BuildBazelRemoteExecutionV2Directory
        include Google::Apis::Core::Hashable
      
        # The subdirectories in the directory.
        # Corresponds to the JSON property `directories`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2DirectoryNode>]
        attr_accessor :directories
      
        # The files in the directory.
        # Corresponds to the JSON property `files`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2FileNode>]
        attr_accessor :files
      
        # The node properties of the Directory.
        # Corresponds to the JSON property `nodeProperties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2NodeProperty>]
        attr_accessor :node_properties
      
        # The symlinks in the directory.
        # Corresponds to the JSON property `symlinks`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2SymlinkNode>]
        attr_accessor :symlinks
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @directories = args[:directories] if args.key?(:directories)
          @files = args[:files] if args.key?(:files)
          @node_properties = args[:node_properties] if args.key?(:node_properties)
          @symlinks = args[:symlinks] if args.key?(:symlinks)
        end
      end
      
      # A `DirectoryNode` represents a child of a Directory which is itself a `
      # Directory` and its associated metadata.
      class BuildBazelRemoteExecutionV2DirectoryNode
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # The name of the directory.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest = args[:digest] if args.key?(:digest)
          @name = args[:name] if args.key?(:name)
        end
      end
      
      # Metadata about an ongoing execution, which will be contained in the metadata
      # field of the Operation.
      class BuildBazelRemoteExecutionV2ExecuteOperationMetadata
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `actionDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :action_digest
      
        # The current stage of execution.
        # Corresponds to the JSON property `stage`
        # @return [String]
        attr_accessor :stage
      
        # If set, the client can use this name with ByteStream.Read to stream the
        # standard error.
        # Corresponds to the JSON property `stderrStreamName`
        # @return [String]
        attr_accessor :stderr_stream_name
      
        # If set, the client can use this name with ByteStream.Read to stream the
        # standard output.
        # Corresponds to the JSON property `stdoutStreamName`
        # @return [String]
        attr_accessor :stdout_stream_name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @action_digest = args[:action_digest] if args.key?(:action_digest)
          @stage = args[:stage] if args.key?(:stage)
          @stderr_stream_name = args[:stderr_stream_name] if args.key?(:stderr_stream_name)
          @stdout_stream_name = args[:stdout_stream_name] if args.key?(:stdout_stream_name)
        end
      end
      
      # A request message for Execution.Execute.
      class BuildBazelRemoteExecutionV2ExecuteRequest
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `actionDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :action_digest
      
        # An `ExecutionPolicy` can be used to control the scheduling of the action.
        # Corresponds to the JSON property `executionPolicy`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutionPolicy]
        attr_accessor :execution_policy
      
        # A `ResultsCachePolicy` is used for fine-grained control over how action
        # outputs are stored in the CAS and Action Cache.
        # Corresponds to the JSON property `resultsCachePolicy`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ResultsCachePolicy]
        attr_accessor :results_cache_policy
      
        # If true, the action will be executed even if its result is already present in
        # the ActionCache. The execution is still allowed to be merged with other in-
        # flight executions of the same action, however - semantically, the service MUST
        # only guarantee that the results of an execution with this field set were not
        # visible before the corresponding execution request was sent. Note that actions
        # from execution requests setting this field set are still eligible to be
        # entered into the action cache upon completion, and services SHOULD overwrite
        # any existing entries that may exist. This allows skip_cache_lookup requests to
        # be used as a mechanism for replacing action cache entries that reference
        # outputs no longer available or that are poisoned in any way. If false, the
        # result may be served from the action cache.
        # Corresponds to the JSON property `skipCacheLookup`
        # @return [Boolean]
        attr_accessor :skip_cache_lookup
        alias_method :skip_cache_lookup?, :skip_cache_lookup
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @action_digest = args[:action_digest] if args.key?(:action_digest)
          @execution_policy = args[:execution_policy] if args.key?(:execution_policy)
          @results_cache_policy = args[:results_cache_policy] if args.key?(:results_cache_policy)
          @skip_cache_lookup = args[:skip_cache_lookup] if args.key?(:skip_cache_lookup)
        end
      end
      
      # The response message for Execution.Execute, which will be contained in the
      # response field of the Operation.
      class BuildBazelRemoteExecutionV2ExecuteResponse
        include Google::Apis::Core::Hashable
      
        # True if the result was served from cache, false if it was executed.
        # Corresponds to the JSON property `cachedResult`
        # @return [Boolean]
        attr_accessor :cached_result
        alias_method :cached_result?, :cached_result
      
        # Freeform informational message with details on the execution of the action
        # that may be displayed to the user upon failure or when requested explicitly.
        # Corresponds to the JSON property `message`
        # @return [String]
        attr_accessor :message
      
        # An ActionResult represents the result of an Action being run.
        # Corresponds to the JSON property `result`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ActionResult]
        attr_accessor :result
      
        # An optional list of additional log outputs the server wishes to provide. A
        # server can use this to return execution-specific logs however it wishes. This
        # is intended primarily to make it easier for users to debug issues that may be
        # outside of the actual job execution, such as by identifying the worker
        # executing the action or by providing logs from the worker's setup phase. The
        # keys SHOULD be human readable so that a client can display them to a user.
        # Corresponds to the JSON property `serverLogs`
        # @return [Hash<String,Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2LogFile>]
        attr_accessor :server_logs
      
        # The `Status` type defines a logical error model that is suitable for different
        # programming environments, including REST APIs and RPC APIs. It is used by [
        # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
        # data: error code, error message, and error details. You can find out more
        # about this error model and how to work with it in the [API Design Guide](https:
        # //cloud.google.com/apis/design/errors).
        # Corresponds to the JSON property `status`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus]
        attr_accessor :status
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @cached_result = args[:cached_result] if args.key?(:cached_result)
          @message = args[:message] if args.key?(:message)
          @result = args[:result] if args.key?(:result)
          @server_logs = args[:server_logs] if args.key?(:server_logs)
          @status = args[:status] if args.key?(:status)
        end
      end
      
      # ExecutedActionMetadata contains details about a completed execution.
      class BuildBazelRemoteExecutionV2ExecutedActionMetadata
        include Google::Apis::Core::Hashable
      
        # When the worker completed executing the action command.
        # Corresponds to the JSON property `executionCompletedTimestamp`
        # @return [String]
        attr_accessor :execution_completed_timestamp
      
        # When the worker started executing the action command.
        # Corresponds to the JSON property `executionStartTimestamp`
        # @return [String]
        attr_accessor :execution_start_timestamp
      
        # When the worker finished fetching action inputs.
        # Corresponds to the JSON property `inputFetchCompletedTimestamp`
        # @return [String]
        attr_accessor :input_fetch_completed_timestamp
      
        # When the worker started fetching action inputs.
        # Corresponds to the JSON property `inputFetchStartTimestamp`
        # @return [String]
        attr_accessor :input_fetch_start_timestamp
      
        # When the worker finished uploading action outputs.
        # Corresponds to the JSON property `outputUploadCompletedTimestamp`
        # @return [String]
        attr_accessor :output_upload_completed_timestamp
      
        # When the worker started uploading action outputs.
        # Corresponds to the JSON property `outputUploadStartTimestamp`
        # @return [String]
        attr_accessor :output_upload_start_timestamp
      
        # When was the action added to the queue.
        # Corresponds to the JSON property `queuedTimestamp`
        # @return [String]
        attr_accessor :queued_timestamp
      
        # The name of the worker which ran the execution.
        # Corresponds to the JSON property `worker`
        # @return [String]
        attr_accessor :worker
      
        # When the worker completed the action, including all stages.
        # Corresponds to the JSON property `workerCompletedTimestamp`
        # @return [String]
        attr_accessor :worker_completed_timestamp
      
        # When the worker received the action.
        # Corresponds to the JSON property `workerStartTimestamp`
        # @return [String]
        attr_accessor :worker_start_timestamp
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @execution_completed_timestamp = args[:execution_completed_timestamp] if args.key?(:execution_completed_timestamp)
          @execution_start_timestamp = args[:execution_start_timestamp] if args.key?(:execution_start_timestamp)
          @input_fetch_completed_timestamp = args[:input_fetch_completed_timestamp] if args.key?(:input_fetch_completed_timestamp)
          @input_fetch_start_timestamp = args[:input_fetch_start_timestamp] if args.key?(:input_fetch_start_timestamp)
          @output_upload_completed_timestamp = args[:output_upload_completed_timestamp] if args.key?(:output_upload_completed_timestamp)
          @output_upload_start_timestamp = args[:output_upload_start_timestamp] if args.key?(:output_upload_start_timestamp)
          @queued_timestamp = args[:queued_timestamp] if args.key?(:queued_timestamp)
          @worker = args[:worker] if args.key?(:worker)
          @worker_completed_timestamp = args[:worker_completed_timestamp] if args.key?(:worker_completed_timestamp)
          @worker_start_timestamp = args[:worker_start_timestamp] if args.key?(:worker_start_timestamp)
        end
      end
      
      # Capabilities of the remote execution system.
      class BuildBazelRemoteExecutionV2ExecutionCapabilities
        include Google::Apis::Core::Hashable
      
        # Remote execution may only support a single digest function.
        # Corresponds to the JSON property `digestFunction`
        # @return [String]
        attr_accessor :digest_function
      
        # Whether remote execution is enabled for the particular server/instance.
        # Corresponds to the JSON property `execEnabled`
        # @return [Boolean]
        attr_accessor :exec_enabled
        alias_method :exec_enabled?, :exec_enabled
      
        # Allowed values for priority in ResultsCachePolicy Used for querying both cache
        # and execution valid priority ranges.
        # Corresponds to the JSON property `executionPriorityCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PriorityCapabilities]
        attr_accessor :execution_priority_capabilities
      
        # Supported node properties.
        # Corresponds to the JSON property `supportedNodeProperties`
        # @return [Array<String>]
        attr_accessor :supported_node_properties
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest_function = args[:digest_function] if args.key?(:digest_function)
          @exec_enabled = args[:exec_enabled] if args.key?(:exec_enabled)
          @execution_priority_capabilities = args[:execution_priority_capabilities] if args.key?(:execution_priority_capabilities)
          @supported_node_properties = args[:supported_node_properties] if args.key?(:supported_node_properties)
        end
      end
      
      # An `ExecutionPolicy` can be used to control the scheduling of the action.
      class BuildBazelRemoteExecutionV2ExecutionPolicy
        include Google::Apis::Core::Hashable
      
        # The priority (relative importance) of this action. Generally, a lower value
        # means that the action should be run sooner than actions having a greater
        # priority value, but the interpretation of a given value is server- dependent.
        # A priority of 0 means the *default* priority. Priorities may be positive or
        # negative, and such actions should run later or sooner than actions having the
        # default priority, respectively. The particular semantics of this field is up
        # to the server. In particular, every server will have their own supported range
        # of priorities, and will decide how these map into scheduling policy.
        # Corresponds to the JSON property `priority`
        # @return [Fixnum]
        attr_accessor :priority
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @priority = args[:priority] if args.key?(:priority)
        end
      end
      
      # A `FileNode` represents a single file and associated metadata.
      class BuildBazelRemoteExecutionV2FileNode
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # True if file is executable, false otherwise.
        # Corresponds to the JSON property `isExecutable`
        # @return [Boolean]
        attr_accessor :is_executable
        alias_method :is_executable?, :is_executable
      
        # The name of the file.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The node properties of the FileNode.
        # Corresponds to the JSON property `nodeProperties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2NodeProperty>]
        attr_accessor :node_properties
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest = args[:digest] if args.key?(:digest)
          @is_executable = args[:is_executable] if args.key?(:is_executable)
          @name = args[:name] if args.key?(:name)
          @node_properties = args[:node_properties] if args.key?(:node_properties)
        end
      end
      
      # A request message for ContentAddressableStorage.FindMissingBlobs.
      class BuildBazelRemoteExecutionV2FindMissingBlobsRequest
        include Google::Apis::Core::Hashable
      
        # A list of the blobs to check.
        # Corresponds to the JSON property `blobDigests`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest>]
        attr_accessor :blob_digests
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @blob_digests = args[:blob_digests] if args.key?(:blob_digests)
        end
      end
      
      # A response message for ContentAddressableStorage.FindMissingBlobs.
      class BuildBazelRemoteExecutionV2FindMissingBlobsResponse
        include Google::Apis::Core::Hashable
      
        # A list of the blobs requested *not* present in the storage.
        # Corresponds to the JSON property `missingBlobDigests`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest>]
        attr_accessor :missing_blob_digests
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @missing_blob_digests = args[:missing_blob_digests] if args.key?(:missing_blob_digests)
        end
      end
      
      # A response message for ContentAddressableStorage.GetTree.
      class BuildBazelRemoteExecutionV2GetTreeResponse
        include Google::Apis::Core::Hashable
      
        # The directories descended from the requested root.
        # Corresponds to the JSON property `directories`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Directory>]
        attr_accessor :directories
      
        # If present, signifies that there are more results which the client can
        # retrieve by passing this as the page_token in a subsequent request. If empty,
        # signifies that this is the last page of results.
        # Corresponds to the JSON property `nextPageToken`
        # @return [String]
        attr_accessor :next_page_token
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @directories = args[:directories] if args.key?(:directories)
          @next_page_token = args[:next_page_token] if args.key?(:next_page_token)
        end
      end
      
      # A `LogFile` is a log stored in the CAS.
      class BuildBazelRemoteExecutionV2LogFile
        include Google::Apis::Core::Hashable
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # This is a hint as to the purpose of the log, and is set to true if the log is
        # human-readable text that can be usefully displayed to a user, and false
        # otherwise. For instance, if a command-line client wishes to print the server
        # logs to the terminal for a failed action, this allows it to avoid displaying a
        # binary file.
        # Corresponds to the JSON property `humanReadable`
        # @return [Boolean]
        attr_accessor :human_readable
        alias_method :human_readable?, :human_readable
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest = args[:digest] if args.key?(:digest)
          @human_readable = args[:human_readable] if args.key?(:human_readable)
        end
      end
      
      # A single property for FileNodes, DirectoryNodes, and SymlinkNodes. The server
      # is responsible for specifying the property `name`s that it accepts. If
      # permitted by the server, the same `name` may occur multiple times.
      class BuildBazelRemoteExecutionV2NodeProperty
        include Google::Apis::Core::Hashable
      
        # The property name.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The property value.
        # Corresponds to the JSON property `value`
        # @return [String]
        attr_accessor :value
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
          @value = args[:value] if args.key?(:value)
        end
      end
      
      # An `OutputDirectory` is the output in an `ActionResult` corresponding to a
      # directory's full contents rather than a single file.
      class BuildBazelRemoteExecutionV2OutputDirectory
        include Google::Apis::Core::Hashable
      
        # The full path of the directory relative to the working directory. The path
        # separator is a forward slash `/`. Since this is a relative path, it MUST NOT
        # begin with a leading forward slash. The empty string value is allowed, and it
        # denotes the entire working directory.
        # Corresponds to the JSON property `path`
        # @return [String]
        attr_accessor :path
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `treeDigest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :tree_digest
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @path = args[:path] if args.key?(:path)
          @tree_digest = args[:tree_digest] if args.key?(:tree_digest)
        end
      end
      
      # An `OutputFile` is similar to a FileNode, but it is used as an output in an `
      # ActionResult`. It allows a full file path rather than only a name.
      class BuildBazelRemoteExecutionV2OutputFile
        include Google::Apis::Core::Hashable
      
        # The contents of the file if inlining was requested. The server SHOULD NOT
        # inline file contents unless requested by the client in the
        # GetActionResultRequest message. The server MAY omit inlining, even if
        # requested, and MUST do so if inlining would cause the response to exceed
        # message size limits.
        # Corresponds to the JSON property `contents`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :contents
      
        # A content digest. A digest for a given blob consists of the size of the blob
        # and its hash. The hash algorithm to use is defined by the server. The size is
        # considered to be an integral part of the digest and cannot be separated. That
        # is, even if the `hash` field is correctly specified but `size_bytes` is not,
        # the server MUST reject the request. The reason for including the size in the
        # digest is as follows: in a great many cases, the server needs to know the size
        # of the blob it is about to work with prior to starting an operation with it,
        # such as flattening Merkle tree structures or streaming it to a worker.
        # Technically, the server could implement a separate metadata store, but this
        # results in a significantly more complicated implementation as opposed to
        # having the client specify the size up-front (or storing the size along with
        # the digest in every message where digests are embedded). This does mean that
        # the API leaks some implementation details of (what we consider to be) a
        # reasonable server implementation, but we consider this to be a worthwhile
        # tradeoff. When a `Digest` is used to refer to a proto message, it always
        # refers to the message in binary encoded form. To ensure consistent hashing,
        # clients and servers MUST ensure that they serialize messages according to the
        # following rules, even if there are alternate valid encodings for the same
        # message: * Fields are serialized in tag order. * There are no unknown fields. *
        # There are no duplicate fields. * Fields are serialized according to the
        # default semantics for their type. Most protocol buffer implementations will
        # always follow these rules when serializing, but care should be taken to avoid
        # shortcuts. For instance, concatenating two messages to merge them may produce
        # duplicate fields.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Digest]
        attr_accessor :digest
      
        # True if file is executable, false otherwise.
        # Corresponds to the JSON property `isExecutable`
        # @return [Boolean]
        attr_accessor :is_executable
        alias_method :is_executable?, :is_executable
      
        # The supported node properties of the OutputFile, if requested by the Action.
        # Corresponds to the JSON property `nodeProperties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2NodeProperty>]
        attr_accessor :node_properties
      
        # The full path of the file relative to the working directory, including the
        # filename. The path separator is a forward slash `/`. Since this is a relative
        # path, it MUST NOT begin with a leading forward slash.
        # Corresponds to the JSON property `path`
        # @return [String]
        attr_accessor :path
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @contents = args[:contents] if args.key?(:contents)
          @digest = args[:digest] if args.key?(:digest)
          @is_executable = args[:is_executable] if args.key?(:is_executable)
          @node_properties = args[:node_properties] if args.key?(:node_properties)
          @path = args[:path] if args.key?(:path)
        end
      end
      
      # An `OutputSymlink` is similar to a Symlink, but it is used as an output in an `
      # ActionResult`. `OutputSymlink` is binary-compatible with `SymlinkNode`.
      class BuildBazelRemoteExecutionV2OutputSymlink
        include Google::Apis::Core::Hashable
      
        # The supported node properties of the OutputSymlink, if requested by the Action.
        # Corresponds to the JSON property `nodeProperties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2NodeProperty>]
        attr_accessor :node_properties
      
        # The full path of the symlink relative to the working directory, including the
        # filename. The path separator is a forward slash `/`. Since this is a relative
        # path, it MUST NOT begin with a leading forward slash.
        # Corresponds to the JSON property `path`
        # @return [String]
        attr_accessor :path
      
        # The target path of the symlink. The path separator is a forward slash `/`. The
        # target path can be relative to the parent directory of the symlink or it can
        # be an absolute path starting with `/`. Support for absolute paths can be
        # checked using the Capabilities API. The canonical form forbids the substrings `
        # /./` and `//` in the target path. `..` components are allowed anywhere in the
        # target path.
        # Corresponds to the JSON property `target`
        # @return [String]
        attr_accessor :target
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @node_properties = args[:node_properties] if args.key?(:node_properties)
          @path = args[:path] if args.key?(:path)
          @target = args[:target] if args.key?(:target)
        end
      end
      
      # A `Platform` is a set of requirements, such as hardware, operating system, or
      # compiler toolchain, for an Action's execution environment. A `Platform` is
      # represented as a series of key-value pairs representing the properties that
      # are required of the platform.
      class BuildBazelRemoteExecutionV2Platform
        include Google::Apis::Core::Hashable
      
        # The properties that make up this platform. In order to ensure that equivalent `
        # Platform`s always hash to the same value, the properties MUST be
        # lexicographically sorted by name, and then by value. Sorting of strings is
        # done by code point, equivalently, by the UTF-8 bytes.
        # Corresponds to the JSON property `properties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PlatformProperty>]
        attr_accessor :properties
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @properties = args[:properties] if args.key?(:properties)
        end
      end
      
      # A single property for the environment. The server is responsible for
      # specifying the property `name`s that it accepts. If an unknown `name` is
      # provided in the requirements for an Action, the server SHOULD reject the
      # execution request. If permitted by the server, the same `name` may occur
      # multiple times. The server is also responsible for specifying the
      # interpretation of property `value`s. For instance, a property describing how
      # much RAM must be available may be interpreted as allowing a worker with 16GB
      # to fulfill a request for 8GB, while a property describing the OS environment
      # on which the action must be performed may require an exact match with the
      # worker's OS. The server MAY use the `value` of one or more properties to
      # determine how it sets up the execution environment, such as by making specific
      # system files available to the worker.
      class BuildBazelRemoteExecutionV2PlatformProperty
        include Google::Apis::Core::Hashable
      
        # The property name.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The property value.
        # Corresponds to the JSON property `value`
        # @return [String]
        attr_accessor :value
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
          @value = args[:value] if args.key?(:value)
        end
      end
      
      # Allowed values for priority in ResultsCachePolicy Used for querying both cache
      # and execution valid priority ranges.
      class BuildBazelRemoteExecutionV2PriorityCapabilities
        include Google::Apis::Core::Hashable
      
        # 
        # Corresponds to the JSON property `priorities`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2PriorityCapabilitiesPriorityRange>]
        attr_accessor :priorities
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @priorities = args[:priorities] if args.key?(:priorities)
        end
      end
      
      # Supported range of priorities, including boundaries.
      class BuildBazelRemoteExecutionV2PriorityCapabilitiesPriorityRange
        include Google::Apis::Core::Hashable
      
        # 
        # Corresponds to the JSON property `maxPriority`
        # @return [Fixnum]
        attr_accessor :max_priority
      
        # 
        # Corresponds to the JSON property `minPriority`
        # @return [Fixnum]
        attr_accessor :min_priority
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @max_priority = args[:max_priority] if args.key?(:max_priority)
          @min_priority = args[:min_priority] if args.key?(:min_priority)
        end
      end
      
      # An optional Metadata to attach to any RPC request to tell the server about an
      # external context of the request. The server may use this for logging or other
      # purposes. To use it, the client attaches the header to the call using the
      # canonical proto serialization: * name: `build.bazel.remote.execution.v2.
      # requestmetadata-bin` * contents: the base64 encoded binary `RequestMetadata`
      # message. Note: the gRPC library serializes binary headers encoded in base 64
      # by default (https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md#
      # requests). Therefore, if the gRPC library is used to pass/retrieve this
      # metadata, the user may ignore the base64 encoding and assume it is simply
      # serialized as a binary message.
      class BuildBazelRemoteExecutionV2RequestMetadata
        include Google::Apis::Core::Hashable
      
        # An identifier that ties multiple requests to the same action. For example,
        # multiple requests to the CAS, Action Cache, and Execution API are used in
        # order to compile foo.cc.
        # Corresponds to the JSON property `actionId`
        # @return [String]
        attr_accessor :action_id
      
        # An identifier to tie multiple tool invocations together. For example, runs of
        # foo_test, bar_test and baz_test on a post-submit of a given patch.
        # Corresponds to the JSON property `correlatedInvocationsId`
        # @return [String]
        attr_accessor :correlated_invocations_id
      
        # Details for the tool used to call the API.
        # Corresponds to the JSON property `toolDetails`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ToolDetails]
        attr_accessor :tool_details
      
        # An identifier that ties multiple actions together to a final result. For
        # example, multiple actions are required to build and run foo_test.
        # Corresponds to the JSON property `toolInvocationId`
        # @return [String]
        attr_accessor :tool_invocation_id
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @action_id = args[:action_id] if args.key?(:action_id)
          @correlated_invocations_id = args[:correlated_invocations_id] if args.key?(:correlated_invocations_id)
          @tool_details = args[:tool_details] if args.key?(:tool_details)
          @tool_invocation_id = args[:tool_invocation_id] if args.key?(:tool_invocation_id)
        end
      end
      
      # A `ResultsCachePolicy` is used for fine-grained control over how action
      # outputs are stored in the CAS and Action Cache.
      class BuildBazelRemoteExecutionV2ResultsCachePolicy
        include Google::Apis::Core::Hashable
      
        # The priority (relative importance) of this content in the overall cache.
        # Generally, a lower value means a longer retention time or other advantage, but
        # the interpretation of a given value is server-dependent. A priority of 0 means
        # a *default* value, decided by the server. The particular semantics of this
        # field is up to the server. In particular, every server will have their own
        # supported range of priorities, and will decide how these map into retention/
        # eviction policy.
        # Corresponds to the JSON property `priority`
        # @return [Fixnum]
        attr_accessor :priority
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @priority = args[:priority] if args.key?(:priority)
        end
      end
      
      # A response message for Capabilities.GetCapabilities.
      class BuildBazelRemoteExecutionV2ServerCapabilities
        include Google::Apis::Core::Hashable
      
        # Capabilities of the remote cache system.
        # Corresponds to the JSON property `cacheCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2CacheCapabilities]
        attr_accessor :cache_capabilities
      
        # The full version of a given tool.
        # Corresponds to the JSON property `deprecatedApiVersion`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer]
        attr_accessor :deprecated_api_version
      
        # Capabilities of the remote execution system.
        # Corresponds to the JSON property `executionCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2ExecutionCapabilities]
        attr_accessor :execution_capabilities
      
        # The full version of a given tool.
        # Corresponds to the JSON property `highApiVersion`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer]
        attr_accessor :high_api_version
      
        # The full version of a given tool.
        # Corresponds to the JSON property `lowApiVersion`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelSemverSemVer]
        attr_accessor :low_api_version
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @cache_capabilities = args[:cache_capabilities] if args.key?(:cache_capabilities)
          @deprecated_api_version = args[:deprecated_api_version] if args.key?(:deprecated_api_version)
          @execution_capabilities = args[:execution_capabilities] if args.key?(:execution_capabilities)
          @high_api_version = args[:high_api_version] if args.key?(:high_api_version)
          @low_api_version = args[:low_api_version] if args.key?(:low_api_version)
        end
      end
      
      # A `SymlinkNode` represents a symbolic link.
      class BuildBazelRemoteExecutionV2SymlinkNode
        include Google::Apis::Core::Hashable
      
        # The name of the symlink.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The node properties of the SymlinkNode.
        # Corresponds to the JSON property `nodeProperties`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2NodeProperty>]
        attr_accessor :node_properties
      
        # The target path of the symlink. The path separator is a forward slash `/`. The
        # target path can be relative to the parent directory of the symlink or it can
        # be an absolute path starting with `/`. Support for absolute paths can be
        # checked using the Capabilities API. The canonical form forbids the substrings `
        # /./` and `//` in the target path. `..` components are allowed anywhere in the
        # target path.
        # Corresponds to the JSON property `target`
        # @return [String]
        attr_accessor :target
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
          @node_properties = args[:node_properties] if args.key?(:node_properties)
          @target = args[:target] if args.key?(:target)
        end
      end
      
      # Details for the tool used to call the API.
      class BuildBazelRemoteExecutionV2ToolDetails
        include Google::Apis::Core::Hashable
      
        # Name of the tool, e.g. bazel.
        # Corresponds to the JSON property `toolName`
        # @return [String]
        attr_accessor :tool_name
      
        # Version of the tool used for the request, e.g. 5.0.3.
        # Corresponds to the JSON property `toolVersion`
        # @return [String]
        attr_accessor :tool_version
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @tool_name = args[:tool_name] if args.key?(:tool_name)
          @tool_version = args[:tool_version] if args.key?(:tool_version)
        end
      end
      
      # A `Tree` contains all the Directory protos in a single directory Merkle tree,
      # compressed into one message.
      class BuildBazelRemoteExecutionV2Tree
        include Google::Apis::Core::Hashable
      
        # All the child directories: the directories referred to by the root and,
        # recursively, all its children. In order to reconstruct the directory tree, the
        # client must take the digests of each of the child directories and then build
        # up a tree starting from the `root`.
        # Corresponds to the JSON property `children`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Directory>]
        attr_accessor :children
      
        # A `Directory` represents a directory node in a file tree, containing zero or
        # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains
        # its name in the directory, either the digest of its content (either a file
        # blob or a `Directory` proto) or a symlink target, as well as possibly some
        # metadata about the file or directory. In order to ensure that two equivalent
        # directory trees hash to the same value, the following restrictions MUST be
        # obeyed when constructing a a `Directory`: * Every child in the directory must
        # have a path of exactly one segment. Multiple levels of directory hierarchy may
        # not be collapsed. * Each child in the directory must have a unique path
        # segment (file name). Note that while the API itself is case-sensitive, the
        # environment where the Action is executed may or may not be case-sensitive.
        # That is, it is legal to call the API with a Directory that has both "Foo" and "
        # foo" as children, but the Action may be rejected by the remote system upon
        # execution. * The files, directories and symlinks in the directory must each be
        # sorted in lexicographical order by path. The path strings must be sorted by
        # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files,
        # directories, and symlinks must be sorted in lexicographical order by property
        # name. A `Directory` that obeys the restrictions is said to be in canonical
        # form. As an example, the following could be used for a file named `bar` and a
        # directory named `foo` with an executable file named `baz` (hashes shortened
        # for readability): ```json // (Directory proto) ` files: [ ` name: "bar",
        # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": "
        # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: "
        # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto
        # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: `
        # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ```
        # Corresponds to the JSON property `root`
        # @return [Google::Apis::RemotebuildexecutionV2::BuildBazelRemoteExecutionV2Directory]
        attr_accessor :root
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @children = args[:children] if args.key?(:children)
          @root = args[:root] if args.key?(:root)
        end
      end
      
      # A request message for WaitExecution.
      class BuildBazelRemoteExecutionV2WaitExecutionRequest
        include Google::Apis::Core::Hashable
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
        end
      end
      
      # The full version of a given tool.
      class BuildBazelSemverSemVer
        include Google::Apis::Core::Hashable
      
        # The major version, e.g 10 for 10.2.3.
        # Corresponds to the JSON property `major`
        # @return [Fixnum]
        attr_accessor :major
      
        # The minor version, e.g. 2 for 10.2.3.
        # Corresponds to the JSON property `minor`
        # @return [Fixnum]
        attr_accessor :minor
      
        # The patch version, e.g 3 for 10.2.3.
        # Corresponds to the JSON property `patch`
        # @return [Fixnum]
        attr_accessor :patch
      
        # The pre-release version. Either this field or major/minor/patch fields must be
        # filled. They are mutually exclusive. Pre-release versions are assumed to be
        # earlier than any released versions.
        # Corresponds to the JSON property `prerelease`
        # @return [String]
        attr_accessor :prerelease
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @major = args[:major] if args.key?(:major)
          @minor = args[:minor] if args.key?(:minor)
          @patch = args[:patch] if args.key?(:patch)
          @prerelease = args[:prerelease] if args.key?(:prerelease)
        end
      end
      
      # CommandDuration contains the various duration metrics tracked when a bot
      # performs a command.
      class GoogleDevtoolsRemotebuildbotCommandDurations
        include Google::Apis::Core::Hashable
      
        # The time spent preparing the command to be run in a Docker container (includes
        # pulling the Docker image, if necessary).
        # Corresponds to the JSON property `dockerPrep`
        # @return [String]
        attr_accessor :docker_prep
      
        # The timestamp when docker preparation begins.
        # Corresponds to the JSON property `dockerPrepStartTime`
        # @return [String]
        attr_accessor :docker_prep_start_time
      
        # The time spent downloading the input files and constructing the working
        # directory.
        # Corresponds to the JSON property `download`
        # @return [String]
        attr_accessor :download
      
        # The timestamp when downloading the input files begins.
        # Corresponds to the JSON property `downloadStartTime`
        # @return [String]
        attr_accessor :download_start_time
      
        # The timestamp when execution begins.
        # Corresponds to the JSON property `execStartTime`
        # @return [String]
        attr_accessor :exec_start_time
      
        # The time spent executing the command (i.e., doing useful work).
        # Corresponds to the JSON property `execution`
        # @return [String]
        attr_accessor :execution
      
        # The timestamp when preparation is done and bot starts downloading files.
        # Corresponds to the JSON property `isoPrepDone`
        # @return [String]
        attr_accessor :iso_prep_done
      
        # The time spent completing the command, in total.
        # Corresponds to the JSON property `overall`
        # @return [String]
        attr_accessor :overall
      
        # The time spent uploading the stdout logs.
        # Corresponds to the JSON property `stdout`
        # @return [String]
        attr_accessor :stdout
      
        # The time spent uploading the output files.
        # Corresponds to the JSON property `upload`
        # @return [String]
        attr_accessor :upload
      
        # The timestamp when uploading the output files begins.
        # Corresponds to the JSON property `uploadStartTime`
        # @return [String]
        attr_accessor :upload_start_time
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @docker_prep = args[:docker_prep] if args.key?(:docker_prep)
          @docker_prep_start_time = args[:docker_prep_start_time] if args.key?(:docker_prep_start_time)
          @download = args[:download] if args.key?(:download)
          @download_start_time = args[:download_start_time] if args.key?(:download_start_time)
          @exec_start_time = args[:exec_start_time] if args.key?(:exec_start_time)
          @execution = args[:execution] if args.key?(:execution)
          @iso_prep_done = args[:iso_prep_done] if args.key?(:iso_prep_done)
          @overall = args[:overall] if args.key?(:overall)
          @stdout = args[:stdout] if args.key?(:stdout)
          @upload = args[:upload] if args.key?(:upload)
          @upload_start_time = args[:upload_start_time] if args.key?(:upload_start_time)
        end
      end
      
      # CommandEvents contains counters for the number of warnings and errors that
      # occurred during the execution of a command.
      class GoogleDevtoolsRemotebuildbotCommandEvents
        include Google::Apis::Core::Hashable
      
        # Indicates whether we are using a cached Docker image (true) or had to pull the
        # Docker image (false) for this command.
        # Corresponds to the JSON property `dockerCacheHit`
        # @return [Boolean]
        attr_accessor :docker_cache_hit
        alias_method :docker_cache_hit?, :docker_cache_hit
      
        # Docker Image name.
        # Corresponds to the JSON property `dockerImageName`
        # @return [String]
        attr_accessor :docker_image_name
      
        # The input cache miss ratio.
        # Corresponds to the JSON property `inputCacheMiss`
        # @return [Float]
        attr_accessor :input_cache_miss
      
        # The number of errors reported.
        # Corresponds to the JSON property `numErrors`
        # @return [Fixnum]
        attr_accessor :num_errors
      
        # The number of warnings reported.
        # Corresponds to the JSON property `numWarnings`
        # @return [Fixnum]
        attr_accessor :num_warnings
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @docker_cache_hit = args[:docker_cache_hit] if args.key?(:docker_cache_hit)
          @docker_image_name = args[:docker_image_name] if args.key?(:docker_image_name)
          @input_cache_miss = args[:input_cache_miss] if args.key?(:input_cache_miss)
          @num_errors = args[:num_errors] if args.key?(:num_errors)
          @num_warnings = args[:num_warnings] if args.key?(:num_warnings)
        end
      end
      
      # The internal status of the command result.
      class GoogleDevtoolsRemotebuildbotCommandStatus
        include Google::Apis::Core::Hashable
      
        # The status code.
        # Corresponds to the JSON property `code`
        # @return [String]
        attr_accessor :code
      
        # The error message.
        # Corresponds to the JSON property `message`
        # @return [String]
        attr_accessor :message
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @code = args[:code] if args.key?(:code)
          @message = args[:message] if args.key?(:message)
        end
      end
      
      # ResourceUsage is the system resource usage of the host machine.
      class GoogleDevtoolsRemotebuildbotResourceUsage
        include Google::Apis::Core::Hashable
      
        # 
        # Corresponds to the JSON property `cpuUsedPercent`
        # @return [Float]
        attr_accessor :cpu_used_percent
      
        # 
        # Corresponds to the JSON property `diskUsage`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildbotResourceUsageStat]
        attr_accessor :disk_usage
      
        # 
        # Corresponds to the JSON property `memoryUsage`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildbotResourceUsageStat]
        attr_accessor :memory_usage
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @cpu_used_percent = args[:cpu_used_percent] if args.key?(:cpu_used_percent)
          @disk_usage = args[:disk_usage] if args.key?(:disk_usage)
          @memory_usage = args[:memory_usage] if args.key?(:memory_usage)
        end
      end
      
      # 
      class GoogleDevtoolsRemotebuildbotResourceUsageStat
        include Google::Apis::Core::Hashable
      
        # 
        # Corresponds to the JSON property `total`
        # @return [Fixnum]
        attr_accessor :total
      
        # 
        # Corresponds to the JSON property `used`
        # @return [Fixnum]
        attr_accessor :used
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @total = args[:total] if args.key?(:total)
          @used = args[:used] if args.key?(:used)
        end
      end
      
      # AcceleratorConfig defines the accelerator cards to attach to the VM.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaAcceleratorConfig
        include Google::Apis::Core::Hashable
      
        # The number of guest accelerator cards exposed to each VM.
        # Corresponds to the JSON property `acceleratorCount`
        # @return [Fixnum]
        attr_accessor :accelerator_count
      
        # The type of accelerator to attach to each VM, e.g. "nvidia-tesla-k80" for
        # nVidia Tesla K80.
        # Corresponds to the JSON property `acceleratorType`
        # @return [String]
        attr_accessor :accelerator_type
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @accelerator_count = args[:accelerator_count] if args.key?(:accelerator_count)
          @accelerator_type = args[:accelerator_type] if args.key?(:accelerator_type)
        end
      end
      
      # Autoscale defines the autoscaling policy of a worker pool.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaAutoscale
        include Google::Apis::Core::Hashable
      
        # The maximal number of workers. Must be equal to or greater than min_size.
        # Corresponds to the JSON property `maxSize`
        # @return [Fixnum]
        attr_accessor :max_size
      
        # The minimal number of workers. Must be greater than 0.
        # Corresponds to the JSON property `minSize`
        # @return [Fixnum]
        attr_accessor :min_size
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @max_size = args[:max_size] if args.key?(:max_size)
          @min_size = args[:min_size] if args.key?(:min_size)
        end
      end
      
      # The request used for `CreateInstance`.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateInstanceRequest
        include Google::Apis::Core::Hashable
      
        # Instance conceptually encapsulates all Remote Build Execution resources for
        # remote builds. An instance consists of storage and compute resources (for
        # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
        # running remote builds. All Remote Build Execution API calls are scoped to an
        # instance.
        # Corresponds to the JSON property `instance`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance]
        attr_accessor :instance
      
        # ID of the created instance. A valid `instance_id` must: be 6-50 characters
        # long, contain only lowercase letters, digits, hyphens and underscores, start
        # with a lowercase letter, and end with a lowercase letter or a digit.
        # Corresponds to the JSON property `instanceId`
        # @return [String]
        attr_accessor :instance_id
      
        # Resource name of the project containing the instance. Format: `projects/[
        # PROJECT_ID]`.
        # Corresponds to the JSON property `parent`
        # @return [String]
        attr_accessor :parent
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @instance = args[:instance] if args.key?(:instance)
          @instance_id = args[:instance_id] if args.key?(:instance_id)
          @parent = args[:parent] if args.key?(:parent)
        end
      end
      
      # The request used for `CreateWorkerPool`.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateWorkerPoolRequest
        include Google::Apis::Core::Hashable
      
        # Resource name of the instance in which to create the new worker pool. Format: `
        # projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
        # Corresponds to the JSON property `parent`
        # @return [String]
        attr_accessor :parent
      
        # ID of the created worker pool. A valid pool ID must: be 6-50 characters long,
        # contain only lowercase letters, digits, hyphens and underscores, start with a
        # lowercase letter, and end with a lowercase letter or a digit.
        # Corresponds to the JSON property `poolId`
        # @return [String]
        attr_accessor :pool_id
      
        # A worker pool resource in the Remote Build Execution API.
        # Corresponds to the JSON property `workerPool`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool]
        attr_accessor :worker_pool
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @parent = args[:parent] if args.key?(:parent)
          @pool_id = args[:pool_id] if args.key?(:pool_id)
          @worker_pool = args[:worker_pool] if args.key?(:worker_pool)
        end
      end
      
      # The request used for `DeleteInstance`.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteInstanceRequest
        include Google::Apis::Core::Hashable
      
        # Name of the instance to delete. Format: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]`.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
        end
      end
      
      # The request used for DeleteWorkerPool.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteWorkerPoolRequest
        include Google::Apis::Core::Hashable
      
        # Name of the worker pool to delete. Format: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]/workerpools/[POOL_ID]`.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
        end
      end
      
      # FeaturePolicy defines features allowed to be used on RBE instances, as well as
      # instance-wide behavior changes that take effect without opt-in or opt-out at
      # usage time.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy
        include Google::Apis::Core::Hashable
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `containerImageSources`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :container_image_sources
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerAddCapabilities`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_add_capabilities
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerChrootPath`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_chroot_path
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerNetwork`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_network
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerPrivileged`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_privileged
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerRunAsRoot`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_run_as_root
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerRuntime`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_runtime
      
        # Defines whether a feature can be used or what values are accepted.
        # Corresponds to the JSON property `dockerSiblingContainers`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
        attr_accessor :docker_sibling_containers
      
        # linux_isolation allows overriding the docker runtime used for containers
        # started on Linux.
        # Corresponds to the JSON property `linuxIsolation`
        # @return [String]
        attr_accessor :linux_isolation
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @container_image_sources = args[:container_image_sources] if args.key?(:container_image_sources)
          @docker_add_capabilities = args[:docker_add_capabilities] if args.key?(:docker_add_capabilities)
          @docker_chroot_path = args[:docker_chroot_path] if args.key?(:docker_chroot_path)
          @docker_network = args[:docker_network] if args.key?(:docker_network)
          @docker_privileged = args[:docker_privileged] if args.key?(:docker_privileged)
          @docker_run_as_root = args[:docker_run_as_root] if args.key?(:docker_run_as_root)
          @docker_runtime = args[:docker_runtime] if args.key?(:docker_runtime)
          @docker_sibling_containers = args[:docker_sibling_containers] if args.key?(:docker_sibling_containers)
          @linux_isolation = args[:linux_isolation] if args.key?(:linux_isolation)
        end
      end
      
      # Defines whether a feature can be used or what values are accepted.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature
        include Google::Apis::Core::Hashable
      
        # A list of acceptable values. Only effective when the policy is `RESTRICTED`.
        # Corresponds to the JSON property `allowedValues`
        # @return [Array<String>]
        attr_accessor :allowed_values
      
        # The policy of the feature.
        # Corresponds to the JSON property `policy`
        # @return [String]
        attr_accessor :policy
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @allowed_values = args[:allowed_values] if args.key?(:allowed_values)
          @policy = args[:policy] if args.key?(:policy)
        end
      end
      
      # The request used for `GetInstance`.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetInstanceRequest
        include Google::Apis::Core::Hashable
      
        # Name of the instance to retrieve. Format: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]`.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
        end
      end
      
      # The request used for GetWorkerPool.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetWorkerPoolRequest
        include Google::Apis::Core::Hashable
      
        # Name of the worker pool to retrieve. Format: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]/workerpools/[POOL_ID]`.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
        end
      end
      
      # Instance conceptually encapsulates all Remote Build Execution resources for
      # remote builds. An instance consists of storage and compute resources (for
      # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
      # running remote builds. All Remote Build Execution API calls are scoped to an
      # instance.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance
        include Google::Apis::Core::Hashable
      
        # FeaturePolicy defines features allowed to be used on RBE instances, as well as
        # instance-wide behavior changes that take effect without opt-in or opt-out at
        # usage time.
        # Corresponds to the JSON property `featurePolicy`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy]
        attr_accessor :feature_policy
      
        # The location is a GCP region. Currently only `us-central1` is supported.
        # Corresponds to the JSON property `location`
        # @return [String]
        attr_accessor :location
      
        # Output only. Whether stack driver logging is enabled for the instance.
        # Corresponds to the JSON property `loggingEnabled`
        # @return [Boolean]
        attr_accessor :logging_enabled
        alias_method :logging_enabled?, :logging_enabled
      
        # Output only. Instance resource name formatted as: `projects/[PROJECT_ID]/
        # instances/[INSTANCE_ID]`. Name should not be populated when creating an
        # instance since it is provided in the `instance_id` field.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # Output only. State of the instance.
        # Corresponds to the JSON property `state`
        # @return [String]
        attr_accessor :state
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @feature_policy = args[:feature_policy] if args.key?(:feature_policy)
          @location = args[:location] if args.key?(:location)
          @logging_enabled = args[:logging_enabled] if args.key?(:logging_enabled)
          @name = args[:name] if args.key?(:name)
          @state = args[:state] if args.key?(:state)
        end
      end
      
      # 
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListInstancesRequest
        include Google::Apis::Core::Hashable
      
        # Resource name of the project. Format: `projects/[PROJECT_ID]`.
        # Corresponds to the JSON property `parent`
        # @return [String]
        attr_accessor :parent
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @parent = args[:parent] if args.key?(:parent)
        end
      end
      
      # 
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListInstancesResponse
        include Google::Apis::Core::Hashable
      
        # The list of instances in a given project.
        # Corresponds to the JSON property `instances`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance>]
        attr_accessor :instances
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @instances = args[:instances] if args.key?(:instances)
        end
      end
      
      # 
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListWorkerPoolsRequest
        include Google::Apis::Core::Hashable
      
        # Optional. A filter expression that filters resources listed in the response.
        # The expression must specify the field name, a comparison operator, and the
        # value that you want to use for filtering. The value must be a string, a number,
        # or a boolean. String values are case-insensitive. The comparison operator
        # must be either `:`, `=`, `!=`, `>`, `>=`, `<=` or `<`. The `:` operator can be
        # used with string fields to match substrings. For non-string fields it is
        # equivalent to the `=` operator. The `:*` comparison can be used to test
        # whether a key has been defined. You can also filter on nested fields. To
        # filter on multiple expressions, you can separate expression using `AND` and `
        # OR` operators, using parentheses to specify precedence. If neither operator is
        # specified, `AND` is assumed. Examples: Include only pools with more than 100
        # reserved workers: `(worker_count > 100) (worker_config.reserved = true)`
        # Include only pools with a certain label or machines of the n1-standard family:
        # `worker_config.labels.key1 : * OR worker_config.machine_type: n1-standard`
        # Corresponds to the JSON property `filter`
        # @return [String]
        attr_accessor :filter
      
        # Resource name of the instance. Format: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]`.
        # Corresponds to the JSON property `parent`
        # @return [String]
        attr_accessor :parent
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @filter = args[:filter] if args.key?(:filter)
          @parent = args[:parent] if args.key?(:parent)
        end
      end
      
      # 
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListWorkerPoolsResponse
        include Google::Apis::Core::Hashable
      
        # The list of worker pools in a given instance.
        # Corresponds to the JSON property `workerPools`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool>]
        attr_accessor :worker_pools
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @worker_pools = args[:worker_pools] if args.key?(:worker_pools)
        end
      end
      
      # SoleTenancyConfig specifies information required to host a pool on STNs.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig
        include Google::Apis::Core::Hashable
      
        # The sole-tenant node type to host the pool's workers on.
        # Corresponds to the JSON property `nodeType`
        # @return [String]
        attr_accessor :node_type
      
        # Zone in which STNs are reserved.
        # Corresponds to the JSON property `nodesZone`
        # @return [String]
        attr_accessor :nodes_zone
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @node_type = args[:node_type] if args.key?(:node_type)
          @nodes_zone = args[:nodes_zone] if args.key?(:nodes_zone)
        end
      end
      
      # The request used for `UpdateInstance`.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateInstanceRequest
        include Google::Apis::Core::Hashable
      
        # Instance conceptually encapsulates all Remote Build Execution resources for
        # remote builds. An instance consists of storage and compute resources (for
        # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
        # running remote builds. All Remote Build Execution API calls are scoped to an
        # instance.
        # Corresponds to the JSON property `instance`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance]
        attr_accessor :instance
      
        # Deprecated, use instance.logging_enabled instead. Whether to enable
        # Stackdriver logging for this instance.
        # Corresponds to the JSON property `loggingEnabled`
        # @return [Boolean]
        attr_accessor :logging_enabled
        alias_method :logging_enabled?, :logging_enabled
      
        # Deprecated, use instance.Name instead. Name of the instance to update. Format:
        # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The update mask applies to instance. For the `FieldMask` definition, see https:
        # //developers.google.com/protocol-buffers/docs/reference/google.protobuf#
        # fieldmask If an empty update_mask is provided, only the non-default valued
        # field in the worker pool field will be updated. Note that in order to update a
        # field to the default value (zero, false, empty string) an explicit update_mask
        # must be provided.
        # Corresponds to the JSON property `updateMask`
        # @return [String]
        attr_accessor :update_mask
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @instance = args[:instance] if args.key?(:instance)
          @logging_enabled = args[:logging_enabled] if args.key?(:logging_enabled)
          @name = args[:name] if args.key?(:name)
          @update_mask = args[:update_mask] if args.key?(:update_mask)
        end
      end
      
      # The request used for UpdateWorkerPool.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateWorkerPoolRequest
        include Google::Apis::Core::Hashable
      
        # The update mask applies to worker_pool. For the `FieldMask` definition, see
        # https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#
        # fieldmask If an empty update_mask is provided, only the non-default valued
        # field in the worker pool field will be updated. Note that in order to update a
        # field to the default value (zero, false, empty string) an explicit update_mask
        # must be provided.
        # Corresponds to the JSON property `updateMask`
        # @return [String]
        attr_accessor :update_mask
      
        # A worker pool resource in the Remote Build Execution API.
        # Corresponds to the JSON property `workerPool`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool]
        attr_accessor :worker_pool
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @update_mask = args[:update_mask] if args.key?(:update_mask)
          @worker_pool = args[:worker_pool] if args.key?(:worker_pool)
        end
      end
      
      # Defines the configuration to be used for creating workers in the worker pool.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig
        include Google::Apis::Core::Hashable
      
        # AcceleratorConfig defines the accelerator cards to attach to the VM.
        # Corresponds to the JSON property `accelerator`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaAcceleratorConfig]
        attr_accessor :accelerator
      
        # Required. Size of the disk attached to the worker, in GB. See https://cloud.
        # google.com/compute/docs/disks/
        # Corresponds to the JSON property `diskSizeGb`
        # @return [Fixnum]
        attr_accessor :disk_size_gb
      
        # Required. Disk Type to use for the worker. See [Storage options](https://cloud.
        # google.com/compute/docs/disks/#introduction). Currently only `pd-standard` and
        # `pd-ssd` are supported.
        # Corresponds to the JSON property `diskType`
        # @return [String]
        attr_accessor :disk_type
      
        # Labels associated with the workers. Label keys and values can be no longer
        # than 63 characters, can only contain lowercase letters, numeric characters,
        # underscores and dashes. International letters are permitted. Label keys must
        # start with a letter. Label values are optional. There can not be more than 64
        # labels per resource.
        # Corresponds to the JSON property `labels`
        # @return [Hash<String,String>]
        attr_accessor :labels
      
        # Required. Machine type of the worker, such as `n1-standard-2`. See https://
        # cloud.google.com/compute/docs/machine-types for a list of supported machine
        # types. Note that `f1-micro` and `g1-small` are not yet supported.
        # Corresponds to the JSON property `machineType`
        # @return [String]
        attr_accessor :machine_type
      
        # The maximum number of actions a worker can execute concurrently.
        # Corresponds to the JSON property `maxConcurrentActions`
        # @return [Fixnum]
        attr_accessor :max_concurrent_actions
      
        # Minimum CPU platform to use when creating the worker. See [CPU Platforms](
        # https://cloud.google.com/compute/docs/cpu-platforms).
        # Corresponds to the JSON property `minCpuPlatform`
        # @return [String]
        attr_accessor :min_cpu_platform
      
        # Determines the type of network access granted to workers. Possible values: - "
        # public": Workers can connect to the public internet. - "private": Workers can
        # only connect to Google APIs and services. - "restricted-private": Workers can
        # only connect to Google APIs that are reachable through `restricted.googleapis.
        # com` (`199.36.153.4/30`).
        # Corresponds to the JSON property `networkAccess`
        # @return [String]
        attr_accessor :network_access
      
        # Determines whether the worker is reserved (equivalent to a Compute Engine on-
        # demand VM and therefore won't be preempted). See [Preemptible VMs](https://
        # cloud.google.com/preemptible-vms/) for more details.
        # Corresponds to the JSON property `reserved`
        # @return [Boolean]
        attr_accessor :reserved
        alias_method :reserved?, :reserved
      
        # SoleTenancyConfig specifies information required to host a pool on STNs.
        # Corresponds to the JSON property `soleTenancy`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig]
        attr_accessor :sole_tenancy
      
        # The name of the image used by each VM.
        # Corresponds to the JSON property `vmImage`
        # @return [String]
        attr_accessor :vm_image
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @accelerator = args[:accelerator] if args.key?(:accelerator)
          @disk_size_gb = args[:disk_size_gb] if args.key?(:disk_size_gb)
          @disk_type = args[:disk_type] if args.key?(:disk_type)
          @labels = args[:labels] if args.key?(:labels)
          @machine_type = args[:machine_type] if args.key?(:machine_type)
          @max_concurrent_actions = args[:max_concurrent_actions] if args.key?(:max_concurrent_actions)
          @min_cpu_platform = args[:min_cpu_platform] if args.key?(:min_cpu_platform)
          @network_access = args[:network_access] if args.key?(:network_access)
          @reserved = args[:reserved] if args.key?(:reserved)
          @sole_tenancy = args[:sole_tenancy] if args.key?(:sole_tenancy)
          @vm_image = args[:vm_image] if args.key?(:vm_image)
        end
      end
      
      # A worker pool resource in the Remote Build Execution API.
      class GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerPool
        include Google::Apis::Core::Hashable
      
        # Autoscale defines the autoscaling policy of a worker pool.
        # Corresponds to the JSON property `autoscale`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaAutoscale]
        attr_accessor :autoscale
      
        # Channel specifies the release channel of the pool.
        # Corresponds to the JSON property `channel`
        # @return [String]
        attr_accessor :channel
      
        # WorkerPool resource name formatted as: `projects/[PROJECT_ID]/instances/[
        # INSTANCE_ID]/workerpools/[POOL_ID]`. name should not be populated when
        # creating a worker pool since it is provided in the `poolId` field.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # Output only. State of the worker pool.
        # Corresponds to the JSON property `state`
        # @return [String]
        attr_accessor :state
      
        # Defines the configuration to be used for creating workers in the worker pool.
        # Corresponds to the JSON property `workerConfig`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig]
        attr_accessor :worker_config
      
        # The desired number of workers in the worker pool. Must be a value between 0
        # and 15000.
        # Corresponds to the JSON property `workerCount`
        # @return [Fixnum]
        attr_accessor :worker_count
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @autoscale = args[:autoscale] if args.key?(:autoscale)
          @channel = args[:channel] if args.key?(:channel)
          @name = args[:name] if args.key?(:name)
          @state = args[:state] if args.key?(:state)
          @worker_config = args[:worker_config] if args.key?(:worker_config)
          @worker_count = args[:worker_count] if args.key?(:worker_count)
        end
      end
      
      # AdminTemp is a prelimiary set of administration tasks. It's called "Temp"
      # because we do not yet know the best way to represent admin tasks; it's
      # possible that this will be entirely replaced in later versions of this API. If
      # this message proves to be sufficient, it will be renamed in the alpha or beta
      # release of this API. This message (suitably marshalled into a protobuf.Any)
      # can be used as the inline_assignment field in a lease; the lease assignment
      # field should simply be `"admin"` in these cases. This message is heavily based
      # on Swarming administration tasks from the LUCI project (http://github.com/luci/
      # luci-py/appengine/swarming).
      class GoogleDevtoolsRemoteworkersV1test2AdminTemp
        include Google::Apis::Core::Hashable
      
        # The argument to the admin action; see `Command` for semantics.
        # Corresponds to the JSON property `arg`
        # @return [String]
        attr_accessor :arg
      
        # The admin action; see `Command` for legal values.
        # Corresponds to the JSON property `command`
        # @return [String]
        attr_accessor :command
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @arg = args[:arg] if args.key?(:arg)
          @command = args[:command] if args.key?(:command)
        end
      end
      
      # Describes a blob of binary content with its digest.
      class GoogleDevtoolsRemoteworkersV1test2Blob
        include Google::Apis::Core::Hashable
      
        # The contents of the blob.
        # Corresponds to the JSON property `contents`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :contents
      
        # The CommandTask and CommandResult messages assume the existence of a service
        # that can serve blobs of content, identified by a hash and size known as a "
        # digest." The method by which these blobs may be retrieved is not specified
        # here, but a model implementation is in the Remote Execution API's "
        # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
        # will virtually always refer to the contents of a file or a directory. The
        # latter is represented by the byte-encoded Directory message.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest]
        attr_accessor :digest
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @contents = args[:contents] if args.key?(:contents)
          @digest = args[:digest] if args.key?(:digest)
        end
      end
      
      # DEPRECATED - use CommandResult instead. Describes the actual outputs from the
      # task.
      class GoogleDevtoolsRemoteworkersV1test2CommandOutputs
        include Google::Apis::Core::Hashable
      
        # exit_code is only fully reliable if the status' code is OK. If the task
        # exceeded its deadline or was cancelled, the process may still produce an exit
        # code as it is cancelled, and this will be populated, but a successful (zero)
        # is unlikely to be correct unless the status code is OK.
        # Corresponds to the JSON property `exitCode`
        # @return [Fixnum]
        attr_accessor :exit_code
      
        # The CommandTask and CommandResult messages assume the existence of a service
        # that can serve blobs of content, identified by a hash and size known as a "
        # digest." The method by which these blobs may be retrieved is not specified
        # here, but a model implementation is in the Remote Execution API's "
        # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
        # will virtually always refer to the contents of a file or a directory. The
        # latter is represented by the byte-encoded Directory message.
        # Corresponds to the JSON property `outputs`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest]
        attr_accessor :outputs
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @exit_code = args[:exit_code] if args.key?(:exit_code)
          @outputs = args[:outputs] if args.key?(:outputs)
        end
      end
      
      # DEPRECATED - use CommandResult instead. Can be used as part of CompleteRequest.
      # metadata, or are part of a more sophisticated message.
      class GoogleDevtoolsRemoteworkersV1test2CommandOverhead
        include Google::Apis::Core::Hashable
      
        # The elapsed time between calling Accept and Complete. The server will also
        # have its own idea of what this should be, but this excludes the overhead of
        # the RPCs and the bot response time.
        # Corresponds to the JSON property `duration`
        # @return [String]
        attr_accessor :duration
      
        # The amount of time *not* spent executing the command (ie uploading/downloading
        # files).
        # Corresponds to the JSON property `overhead`
        # @return [String]
        attr_accessor :overhead
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @duration = args[:duration] if args.key?(:duration)
          @overhead = args[:overhead] if args.key?(:overhead)
        end
      end
      
      # All information about the execution of a command, suitable for providing as
      # the Bots interface's `Lease.result` field.
      class GoogleDevtoolsRemoteworkersV1test2CommandResult
        include Google::Apis::Core::Hashable
      
        # The elapsed time between calling Accept and Complete. The server will also
        # have its own idea of what this should be, but this excludes the overhead of
        # the RPCs and the bot response time.
        # Corresponds to the JSON property `duration`
        # @return [String]
        attr_accessor :duration
      
        # The exit code of the process. An exit code of "0" should only be trusted if `
        # status` has a code of OK (otherwise it may simply be unset).
        # Corresponds to the JSON property `exitCode`
        # @return [Fixnum]
        attr_accessor :exit_code
      
        # Implementation-dependent metadata about the task. Both servers and bots may
        # define messages which can be encoded here; bots are free to provide metadata
        # in multiple formats, and servers are free to choose one or more of the values
        # to process and ignore others. In particular, it is *not* considered an error
        # for the bot to provide the server with a field that it doesn't know about.
        # Corresponds to the JSON property `metadata`
        # @return [Array<Hash<String,Object>>]
        attr_accessor :metadata
      
        # The CommandTask and CommandResult messages assume the existence of a service
        # that can serve blobs of content, identified by a hash and size known as a "
        # digest." The method by which these blobs may be retrieved is not specified
        # here, but a model implementation is in the Remote Execution API's "
        # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
        # will virtually always refer to the contents of a file or a directory. The
        # latter is represented by the byte-encoded Directory message.
        # Corresponds to the JSON property `outputs`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest]
        attr_accessor :outputs
      
        # The amount of time *not* spent executing the command (ie uploading/downloading
        # files).
        # Corresponds to the JSON property `overhead`
        # @return [String]
        attr_accessor :overhead
      
        # The `Status` type defines a logical error model that is suitable for different
        # programming environments, including REST APIs and RPC APIs. It is used by [
        # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
        # data: error code, error message, and error details. You can find out more
        # about this error model and how to work with it in the [API Design Guide](https:
        # //cloud.google.com/apis/design/errors).
        # Corresponds to the JSON property `status`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus]
        attr_accessor :status
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @duration = args[:duration] if args.key?(:duration)
          @exit_code = args[:exit_code] if args.key?(:exit_code)
          @metadata = args[:metadata] if args.key?(:metadata)
          @outputs = args[:outputs] if args.key?(:outputs)
          @overhead = args[:overhead] if args.key?(:overhead)
          @status = args[:status] if args.key?(:status)
        end
      end
      
      # Describes a shell-style task to execute, suitable for providing as the Bots
      # interface's `Lease.payload` field.
      class GoogleDevtoolsRemoteworkersV1test2CommandTask
        include Google::Apis::Core::Hashable
      
        # Describes the expected outputs of the command.
        # Corresponds to the JSON property `expectedOutputs`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskOutputs]
        attr_accessor :expected_outputs
      
        # Describes the inputs to a shell-style task.
        # Corresponds to the JSON property `inputs`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskInputs]
        attr_accessor :inputs
      
        # Describes the timeouts associated with this task.
        # Corresponds to the JSON property `timeouts`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskTimeouts]
        attr_accessor :timeouts
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @expected_outputs = args[:expected_outputs] if args.key?(:expected_outputs)
          @inputs = args[:inputs] if args.key?(:inputs)
          @timeouts = args[:timeouts] if args.key?(:timeouts)
        end
      end
      
      # Describes the inputs to a shell-style task.
      class GoogleDevtoolsRemoteworkersV1test2CommandTaskInputs
        include Google::Apis::Core::Hashable
      
        # The command itself to run (e.g., argv). This field should be passed directly
        # to the underlying operating system, and so it must be sensible to that
        # operating system. For example, on Windows, the first argument might be "C:\
        # Windows\System32\ping.exe" - that is, using drive letters and backslashes. A
        # command for a *nix system, on the other hand, would use forward slashes. All
        # other fields in the RWAPI must consistently use forward slashes, since those
        # fields may be interpretted by both the service and the bot.
        # Corresponds to the JSON property `arguments`
        # @return [Array<String>]
        attr_accessor :arguments
      
        # All environment variables required by the task.
        # Corresponds to the JSON property `environmentVariables`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2CommandTaskInputsEnvironmentVariable>]
        attr_accessor :environment_variables
      
        # The input filesystem to be set up prior to the task beginning. The contents
        # should be a repeated set of FileMetadata messages though other formats are
        # allowed if better for the implementation (eg, a LUCI-style .isolated file).
        # This field is repeated since implementations might want to cache the metadata,
        # in which case it may be useful to break up portions of the filesystem that
        # change frequently (eg, specific input files) from those that don't (eg,
        # standard header files).
        # Corresponds to the JSON property `files`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest>]
        attr_accessor :files
      
        # Inline contents for blobs expected to be needed by the bot to execute the task.
        # For example, contents of entries in `files` or blobs that are indirectly
        # referenced by an entry there. The bot should check against this list before
        # downloading required task inputs to reduce the number of communications
        # between itself and the remote CAS server.
        # Corresponds to the JSON property `inlineBlobs`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Blob>]
        attr_accessor :inline_blobs
      
        # Directory from which a command is executed. It is a relative directory with
        # respect to the bot's working directory (i.e., "./"). If it is non-empty, then
        # it must exist under "./". Otherwise, "./" will be used.
        # Corresponds to the JSON property `workingDirectory`
        # @return [String]
        attr_accessor :working_directory
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @arguments = args[:arguments] if args.key?(:arguments)
          @environment_variables = args[:environment_variables] if args.key?(:environment_variables)
          @files = args[:files] if args.key?(:files)
          @inline_blobs = args[:inline_blobs] if args.key?(:inline_blobs)
          @working_directory = args[:working_directory] if args.key?(:working_directory)
        end
      end
      
      # An environment variable required by this task.
      class GoogleDevtoolsRemoteworkersV1test2CommandTaskInputsEnvironmentVariable
        include Google::Apis::Core::Hashable
      
        # The envvar name.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The envvar value.
        # Corresponds to the JSON property `value`
        # @return [String]
        attr_accessor :value
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @name = args[:name] if args.key?(:name)
          @value = args[:value] if args.key?(:value)
        end
      end
      
      # Describes the expected outputs of the command.
      class GoogleDevtoolsRemoteworkersV1test2CommandTaskOutputs
        include Google::Apis::Core::Hashable
      
        # A list of expected directories, relative to the execution root. All paths MUST
        # be delimited by forward slashes.
        # Corresponds to the JSON property `directories`
        # @return [Array<String>]
        attr_accessor :directories
      
        # A list of expected files, relative to the execution root. All paths MUST be
        # delimited by forward slashes.
        # Corresponds to the JSON property `files`
        # @return [Array<String>]
        attr_accessor :files
      
        # The destination to which any stderr should be sent. The method by which the
        # bot should send the stream contents to that destination is not defined in this
        # API. As examples, the destination could be a file referenced in the `files`
        # field in this message, or it could be a URI that must be written via the
        # ByteStream API.
        # Corresponds to the JSON property `stderrDestination`
        # @return [String]
        attr_accessor :stderr_destination
      
        # The destination to which any stdout should be sent. The method by which the
        # bot should send the stream contents to that destination is not defined in this
        # API. As examples, the destination could be a file referenced in the `files`
        # field in this message, or it could be a URI that must be written via the
        # ByteStream API.
        # Corresponds to the JSON property `stdoutDestination`
        # @return [String]
        attr_accessor :stdout_destination
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @directories = args[:directories] if args.key?(:directories)
          @files = args[:files] if args.key?(:files)
          @stderr_destination = args[:stderr_destination] if args.key?(:stderr_destination)
          @stdout_destination = args[:stdout_destination] if args.key?(:stdout_destination)
        end
      end
      
      # Describes the timeouts associated with this task.
      class GoogleDevtoolsRemoteworkersV1test2CommandTaskTimeouts
        include Google::Apis::Core::Hashable
      
        # This specifies the maximum time that the task can run, excluding the time
        # required to download inputs or upload outputs. That is, the worker will
        # terminate the task if it runs longer than this.
        # Corresponds to the JSON property `execution`
        # @return [String]
        attr_accessor :execution
      
        # This specifies the maximum amount of time the task can be idle - that is, go
        # without generating some output in either stdout or stderr. If the process is
        # silent for more than the specified time, the worker will terminate the task.
        # Corresponds to the JSON property `idle`
        # @return [String]
        attr_accessor :idle
      
        # If the execution or IO timeouts are exceeded, the worker will try to
        # gracefully terminate the task and return any existing logs. However, tasks may
        # be hard-frozen in which case this process will fail. This timeout specifies
        # how long to wait for a terminated task to shut down gracefully (e.g. via
        # SIGTERM) before we bring down the hammer (e.g. SIGKILL on *nix,
        # CTRL_BREAK_EVENT on Windows).
        # Corresponds to the JSON property `shutdown`
        # @return [String]
        attr_accessor :shutdown
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @execution = args[:execution] if args.key?(:execution)
          @idle = args[:idle] if args.key?(:idle)
          @shutdown = args[:shutdown] if args.key?(:shutdown)
        end
      end
      
      # The CommandTask and CommandResult messages assume the existence of a service
      # that can serve blobs of content, identified by a hash and size known as a "
      # digest." The method by which these blobs may be retrieved is not specified
      # here, but a model implementation is in the Remote Execution API's "
      # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
      # will virtually always refer to the contents of a file or a directory. The
      # latter is represented by the byte-encoded Directory message.
      class GoogleDevtoolsRemoteworkersV1test2Digest
        include Google::Apis::Core::Hashable
      
        # A string-encoded hash (eg "1a2b3c", not the byte array [0x1a, 0x2b, 0x3c])
        # using an implementation-defined hash algorithm (eg SHA-256).
        # Corresponds to the JSON property `hash`
        # @return [String]
        attr_accessor :hash_prop
      
        # The size of the contents. While this is not strictly required as part of an
        # identifier (after all, any given hash will have exactly one canonical size),
        # it's useful in almost all cases when one might want to send or retrieve blobs
        # of content and is included here for this reason.
        # Corresponds to the JSON property `sizeBytes`
        # @return [Fixnum]
        attr_accessor :size_bytes
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @hash_prop = args[:hash_prop] if args.key?(:hash_prop)
          @size_bytes = args[:size_bytes] if args.key?(:size_bytes)
        end
      end
      
      # The contents of a directory. Similar to the equivalent message in the Remote
      # Execution API.
      class GoogleDevtoolsRemoteworkersV1test2Directory
        include Google::Apis::Core::Hashable
      
        # Any subdirectories
        # Corresponds to the JSON property `directories`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2DirectoryMetadata>]
        attr_accessor :directories
      
        # The files in this directory
        # Corresponds to the JSON property `files`
        # @return [Array<Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2FileMetadata>]
        attr_accessor :files
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @directories = args[:directories] if args.key?(:directories)
          @files = args[:files] if args.key?(:files)
        end
      end
      
      # The metadata for a directory. Similar to the equivalent message in the Remote
      # Execution API.
      class GoogleDevtoolsRemoteworkersV1test2DirectoryMetadata
        include Google::Apis::Core::Hashable
      
        # The CommandTask and CommandResult messages assume the existence of a service
        # that can serve blobs of content, identified by a hash and size known as a "
        # digest." The method by which these blobs may be retrieved is not specified
        # here, but a model implementation is in the Remote Execution API's "
        # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
        # will virtually always refer to the contents of a file or a directory. The
        # latter is represented by the byte-encoded Directory message.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest]
        attr_accessor :digest
      
        # The path of the directory, as in FileMetadata.path.
        # Corresponds to the JSON property `path`
        # @return [String]
        attr_accessor :path
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @digest = args[:digest] if args.key?(:digest)
          @path = args[:path] if args.key?(:path)
        end
      end
      
      # The metadata for a file. Similar to the equivalent message in the Remote
      # Execution API.
      class GoogleDevtoolsRemoteworkersV1test2FileMetadata
        include Google::Apis::Core::Hashable
      
        # If the file is small enough, its contents may also or alternatively be listed
        # here.
        # Corresponds to the JSON property `contents`
        # NOTE: Values are automatically base64 encoded/decoded in the client library.
        # @return [String]
        attr_accessor :contents
      
        # The CommandTask and CommandResult messages assume the existence of a service
        # that can serve blobs of content, identified by a hash and size known as a "
        # digest." The method by which these blobs may be retrieved is not specified
        # here, but a model implementation is in the Remote Execution API's "
        # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
        # will virtually always refer to the contents of a file or a directory. The
        # latter is represented by the byte-encoded Directory message.
        # Corresponds to the JSON property `digest`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleDevtoolsRemoteworkersV1test2Digest]
        attr_accessor :digest
      
        # Properties of the file
        # Corresponds to the JSON property `isExecutable`
        # @return [Boolean]
        attr_accessor :is_executable
        alias_method :is_executable?, :is_executable
      
        # The path of this file. If this message is part of the CommandOutputs.outputs
        # fields, the path is relative to the execution root and must correspond to an
        # entry in CommandTask.outputs.files. If this message is part of a Directory
        # message, then the path is relative to the root of that directory. All paths
        # MUST be delimited by forward slashes.
        # Corresponds to the JSON property `path`
        # @return [String]
        attr_accessor :path
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @contents = args[:contents] if args.key?(:contents)
          @digest = args[:digest] if args.key?(:digest)
          @is_executable = args[:is_executable] if args.key?(:is_executable)
          @path = args[:path] if args.key?(:path)
        end
      end
      
      # This resource represents a long-running operation that is the result of a
      # network API call.
      class GoogleLongrunningOperation
        include Google::Apis::Core::Hashable
      
        # If the value is `false`, it means the operation is still in progress. If `true`
        # , the operation is completed, and either `error` or `response` is available.
        # Corresponds to the JSON property `done`
        # @return [Boolean]
        attr_accessor :done
        alias_method :done?, :done
      
        # The `Status` type defines a logical error model that is suitable for different
        # programming environments, including REST APIs and RPC APIs. It is used by [
        # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
        # data: error code, error message, and error details. You can find out more
        # about this error model and how to work with it in the [API Design Guide](https:
        # //cloud.google.com/apis/design/errors).
        # Corresponds to the JSON property `error`
        # @return [Google::Apis::RemotebuildexecutionV2::GoogleRpcStatus]
        attr_accessor :error
      
        # Service-specific metadata associated with the operation. It typically contains
        # progress information and common metadata such as create time. Some services
        # might not provide such metadata. Any method that returns a long-running
        # operation should document the metadata type, if any.
        # Corresponds to the JSON property `metadata`
        # @return [Hash<String,Object>]
        attr_accessor :metadata
      
        # The server-assigned name, which is only unique within the same service that
        # originally returns it. If you use the default HTTP mapping, the `name` should
        # be a resource name ending with `operations/`unique_id``.
        # Corresponds to the JSON property `name`
        # @return [String]
        attr_accessor :name
      
        # The normal response of the operation in case of success. If the original
        # method returns no data on success, such as `Delete`, the response is `google.
        # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`,
        # the response should be the resource. For other methods, the response should
        # have the type `XxxResponse`, where `Xxx` is the original method name. For
        # example, if the original method name is `TakeSnapshot()`, the inferred
        # response type is `TakeSnapshotResponse`.
        # Corresponds to the JSON property `response`
        # @return [Hash<String,Object>]
        attr_accessor :response
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @done = args[:done] if args.key?(:done)
          @error = args[:error] if args.key?(:error)
          @metadata = args[:metadata] if args.key?(:metadata)
          @name = args[:name] if args.key?(:name)
          @response = args[:response] if args.key?(:response)
        end
      end
      
      # The `Status` type defines a logical error model that is suitable for different
      # programming environments, including REST APIs and RPC APIs. It is used by [
      # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
      # data: error code, error message, and error details. You can find out more
      # about this error model and how to work with it in the [API Design Guide](https:
      # //cloud.google.com/apis/design/errors).
      class GoogleRpcStatus
        include Google::Apis::Core::Hashable
      
        # The status code, which should be an enum value of google.rpc.Code.
        # Corresponds to the JSON property `code`
        # @return [Fixnum]
        attr_accessor :code
      
        # A list of messages that carry the error details. There is a common set of
        # message types for APIs to use.
        # Corresponds to the JSON property `details`
        # @return [Array<Hash<String,Object>>]
        attr_accessor :details
      
        # A developer-facing error message, which should be in English. Any user-facing
        # error message should be localized and sent in the google.rpc.Status.details
        # field, or localized by the client.
        # Corresponds to the JSON property `message`
        # @return [String]
        attr_accessor :message
      
        def initialize(**args)
           update!(**args)
        end
      
        # Update properties of this object
        def update!(**args)
          @code = args[:code] if args.key?(:code)
          @details = args[:details] if args.key?(:details)
          @message = args[:message] if args.key?(:message)
        end
      end
    end
  end
end