# Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. require 'date' require 'google/apis/core/base_service' require 'google/apis/core/json_representation' require 'google/apis/core/hashable' require 'google/apis/errors' module Google module Apis module SpannerV1 # A backup of a Cloud Spanner database. class Backup include Google::Apis::Core::Hashable # Output only. The backup will contain an externally consistent # copy of the database at the timestamp specified by # `create_time`. `create_time` is approximately the time the # CreateBackup request is received. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Required for the CreateBackup operation. # Name of the database from which this backup was # created. This needs to be in the same instance as the backup. # Values are of the form # `projects//instances//databases/`. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # Required for the CreateBackup # operation. The expiration time of the backup, with microseconds # granularity that must be at least 6 hours and at most 366 days # from the time the CreateBackup request is processed. Once the `expire_time` # has passed, the backup is eligible to be automatically deleted by Cloud # Spanner to free the resources used by the backup. # Corresponds to the JSON property `expireTime` # @return [String] attr_accessor :expire_time # Output only for the CreateBackup operation. # Required for the UpdateBackup operation. # A globally unique identifier for the backup which cannot be # changed. Values are of the form # `projects//instances//backups/a-z*[a-z0-9]` # The final segment of the name must be between 2 and 60 characters # in length. # The backup is stored in the location(s) specified in the instance # configuration of the instance containing the backup, identified # by the prefix of the backup name of the form # `projects//instances/`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Output only. The names of the restored databases that reference the backup. # The database names are of # the form `projects//instances//databases/`. # Referencing databases may exist in different instances. The existence of # any referencing database prevents the backup from being deleted. When a # restored database from the backup enters the `READY` state, the reference # to the backup is removed. # Corresponds to the JSON property `referencingDatabases` # @return [Array] attr_accessor :referencing_databases # Output only. Size of the backup in bytes. # Corresponds to the JSON property `sizeBytes` # @return [Fixnum] attr_accessor :size_bytes # Output only. The current state of the backup. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_time = args[:create_time] if args.key?(:create_time) @database = args[:database] if args.key?(:database) @expire_time = args[:expire_time] if args.key?(:expire_time) @name = args[:name] if args.key?(:name) @referencing_databases = args[:referencing_databases] if args.key?(:referencing_databases) @size_bytes = args[:size_bytes] if args.key?(:size_bytes) @state = args[:state] if args.key?(:state) end end # Information about a backup. class BackupInfo include Google::Apis::Core::Hashable # Name of the backup. # Corresponds to the JSON property `backup` # @return [String] attr_accessor :backup # The backup contains an externally consistent copy of `source_database` at # the timestamp specified by `create_time`. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Name of the database the backup was created from. # Corresponds to the JSON property `sourceDatabase` # @return [String] attr_accessor :source_database def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup = args[:backup] if args.key?(:backup) @create_time = args[:create_time] if args.key?(:create_time) @source_database = args[:source_database] if args.key?(:source_database) end end # The request for BatchCreateSessions. class BatchCreateSessionsRequest include Google::Apis::Core::Hashable # Required. The number of sessions to be created in this batch call. # The API may return fewer than the requested number of sessions. If a # specific number of sessions are desired, the client can make additional # calls to BatchCreateSessions (adjusting # session_count as necessary). # Corresponds to the JSON property `sessionCount` # @return [Fixnum] attr_accessor :session_count # A session in the Cloud Spanner API. # Corresponds to the JSON property `sessionTemplate` # @return [Google::Apis::SpannerV1::Session] attr_accessor :session_template def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session_count = args[:session_count] if args.key?(:session_count) @session_template = args[:session_template] if args.key?(:session_template) end end # The response for BatchCreateSessions. class BatchCreateSessionsResponse include Google::Apis::Core::Hashable # The freshly created sessions. # Corresponds to the JSON property `session` # @return [Array] attr_accessor :session def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session = args[:session] if args.key?(:session) end end # The request for BeginTransaction. class BeginTransactionRequest include Google::Apis::Core::Hashable # # Transactions # Each session can have at most one active transaction at a time (note that # standalone reads and queries use a transaction internally and do count # towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. # It is not necessary to create a new session for each transaction. # # Transaction Modes # Cloud Spanner supports three transaction modes: # 1. Locking read-write. This type of transaction is the only way # to write data into Cloud Spanner. These transactions rely on # pessimistic locking and, if necessary, two-phase commit. # Locking read-write transactions may abort, requiring the # application to retry. # 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow # writes. Snapshot read-only transactions can be configured to # read at timestamps in the past. Snapshot read-only # transactions do not need to be committed. # 3. Partitioned DML. This type of transaction is used to execute # a single Partitioned DML statement. Partitioned DML partitions # the key space and runs the DML statement over each partition # in parallel using separate, internal transactions that commit # independently. Partitioned DML transactions do not need to be # committed. # For transactions that only read, snapshot read-only transactions # provide simpler semantics and are almost always faster. In # particular, read-only transactions do not take locks, so they do # not conflict with read-write transactions. As a consequence of not # taking locks, they also do not abort, so retry loops are not needed. # Transactions may only read/write data in a single database. They # may, however, read/write data in different tables within that # database. # ## Locking Read-Write Transactions # Locking transactions may be used to atomically read-modify-write # data anywhere in a database. This type of transaction is externally # consistent. # Clients should attempt to minimize the amount of time a transaction # is active. Faster transactions commit with higher probability # and cause less contention. Cloud Spanner attempts to keep read locks # active as long as the transaction continues to do reads, and the # transaction has not been terminated by # Commit or # Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a # transaction's locks and abort it. # Conceptually, a read-write transaction consists of zero or more # reads or SQL statements followed by # Commit. At any time before # Commit, the client can send a # Rollback request to abort the # transaction. # ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired # are still valid at commit time, and it is able to acquire write # locks for all writes. Cloud Spanner can abort the transaction for any # reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees # that the transaction has not modified any user data in Cloud Spanner. # Unless the transaction commits, Cloud Spanner makes no guarantees about # how long the transaction's locks were held for. It is an error to # use Cloud Spanner locks for any sort of mutual exclusion other than # between Cloud Spanner transactions themselves. # ### Retrying Aborted Transactions # When a transaction aborts, the application can choose to retry the # whole transaction again. To maximize the chances of successfully # committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock # priority increases with each consecutive abort, meaning that each # attempt has a slightly better chance of success than the previous. # Under some circumstances (e.g., many transactions attempting to # modify the same row(s)), a transaction can abort many times in a # short period before successfully committing. Thus, it is not a good # idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent # retrying. # ### Idle Transactions # A transaction is considered idle if it has no outstanding reads or # SQL queries and has not started a read or SQL query within the last 10 # seconds. Idle transactions can be aborted by Cloud Spanner so that they # don't hold on to locks indefinitely. In that case, the commit will # fail with error `ABORTED`. # If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the # transaction from becoming idle. # ## Snapshot Read-Only Transactions # Snapshot read-only transactions provides a simpler method than # locking read-write transactions for doing several consistent # reads. However, this type of transaction does not support writes. # Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that # timestamp. Since they do not acquire locks, they do not block # concurrent read-write transactions. # Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read # timestamp is garbage collected; however, the default garbage # collection policy is generous enough that most applications do not # need to worry about this in practice. # Snapshot read-only transactions do not need to call # Commit or # Rollback (and in fact are not # permitted to do so). # To execute a snapshot transaction, the client specifies a timestamp # bound, which tells Cloud Spanner how to choose a read timestamp. # The types of timestamp bound are: # - Strong (the default). # - Bounded staleness. # - Exact staleness. # If the Cloud Spanner database to be read is geographically distributed, # stale read-only transactions can execute more quickly than strong # or read-write transaction, because they are able to execute far # from the leader replica. # Each type of timestamp bound is discussed in detail below. # ### Strong # Strong reads are guaranteed to see the effects of all transactions # that have committed before the start of the read. Furthermore, all # rows yielded by a single read are consistent with each other -- if # any part of the read observes a transaction, all parts of the read # see the transaction. # Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are # concurrent writes. If consistency across reads is required, the # reads should be executed within a transaction or at an exact read # timestamp. # See TransactionOptions.ReadOnly.strong. # ### Exact Staleness # These timestamp bounds execute reads at a user-specified # timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe # modifications done by all transactions with a commit timestamp <= # the read timestamp, and observe none of the modifications done by # transactions with a larger commit timestamp. They will block until # all conflicting transactions that may be assigned commit timestamps # <= the read timestamp have finished. # The timestamp can either be expressed as an absolute Cloud Spanner commit # timestamp or a staleness relative to the current time. # These modes do not require a "negotiation phase" to pick a # timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, # boundedly stale reads usually return fresher results. # See TransactionOptions.ReadOnly.read_timestamp and # TransactionOptions.ReadOnly.exact_staleness. # ### Bounded Staleness # Bounded staleness modes allow Cloud Spanner to pick the read timestamp, # subject to a user-provided staleness bound. Cloud Spanner chooses the # newest timestamp within the staleness bound that allows execution # of the reads at the closest available replica without blocking. # All rows yielded are consistent with each other -- if any part of # the read observes a transaction, all parts of the read see the # transaction. Boundedly stale reads are not repeatable: two stale # reads, even if they use the same staleness bound, can execute at # different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase # negotiates a timestamp among all replicas needed to serve the # read. In the second phase, reads are executed at the negotiated # timestamp. # As a result of the two phase execution, bounded staleness reads are # usually a little slower than comparable exact staleness # reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. # Because the timestamp negotiation requires up-front knowledge of # which rows will be read, it can only be used with single-use # read-only transactions. # See TransactionOptions.ReadOnly.max_staleness and # TransactionOptions.ReadOnly.min_read_timestamp. # ### Old Read Timestamps and Garbage Collection # Cloud Spanner continuously garbage collects deleted and overwritten data # in the background to reclaim storage space. This process is known # as "version GC". By default, version GC reclaims versions after they # are one hour old. Because of this, Cloud Spanner cannot perform reads # at read timestamps more than one hour in the past. This # restriction also applies to in-progress reads and/or SQL queries whose # timestamp become too old while executing. Reads and SQL queries with # too-old read timestamps fail with the error `FAILED_PRECONDITION`. # ## Partitioned DML Transactions # Partitioned DML transactions are used to execute DML statements with a # different execution strategy that provides different, and often better, # scalability properties for large, table-wide operations than DML in a # ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, # should prefer using ReadWrite transactions. # Partitioned DML partitions the keyspace and runs the DML statement on each # partition in separate, internal transactions. These transactions commit # automatically when complete, and run independently from one another. # To reduce lock contention, this execution strategy only acquires read locks # on rows that match the WHERE clause of the statement. Additionally, the # smaller per-partition transactions hold locks for less time. # That said, Partitioned DML is not a drop-in replacement for standard DML used # in ReadWrite transactions. # - The DML statement must be fully-partitionable. Specifically, the statement # must be expressible as the union of many statements which each access only # a single row of the table. # - The statement is not applied atomically to all rows of the table. Rather, # the statement is applied atomically to partitions of the table, in # independent transactions. Secondary index rows are updated atomically # with the base table rows. # - Partitioned DML does not guarantee exactly-once execution semantics # against a partition. The statement will be applied at least once to each # partition. It is strongly recommended that the DML statement should be # idempotent to avoid unexpected results. For instance, it is potentially # dangerous to run a statement such as # `UPDATE table SET column = column + 1` as it could be run multiple times # against some rows. # - The partitions are committed automatically - there is no support for # Commit or Rollback. If the call returns an error, or if the client issuing # the ExecuteSql call dies, it is possible that some rows had the statement # executed on them successfully. It is also possible that statement was # never executed against other rows. # - Partitioned DML transactions may only contain the execution of a single # DML statement via ExecuteSql or ExecuteStreamingSql. # - If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the # operation is stopped at that point and an error is returned. It is # possible that at this point, some partitions have been committed (or even # committed multiple times), and other partitions have not been run at all. # Given the above, Partitioned DML is good fit for large, database-wide, # operations that are idempotent, such as deleting old rows from a very large # table. # Corresponds to the JSON property `options` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :options def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @options = args[:options] if args.key?(:options) end end # Associates `members` with a `role`. class Binding include Google::Apis::Core::Hashable # Represents a textual expression in the Common Expression Language (CEL) # syntax. CEL is a C-like expression language. The syntax and semantics of CEL # are documented at https://github.com/google/cel-spec. # Example (Comparison): # title: "Summary size limit" # description: "Determines if a summary is less than 100 chars" # expression: "document.summary.size() < 100" # Example (Equality): # title: "Requestor is owner" # description: "Determines if requestor is the document owner" # expression: "document.owner == request.auth.claims.email" # Example (Logic): # title: "Public documents" # description: "Determine whether the document should be publicly visible" # expression: "document.type != 'private' && document.type != 'internal'" # Example (Data Manipulation): # title: "Notification string" # description: "Create a notification string with a timestamp." # expression: "'New message received at ' + string(document.create_time)" # The exact variables and functions that may be referenced within an expression # are determined by the service that evaluates it. See the service # documentation for additional information. # Corresponds to the JSON property `condition` # @return [Google::Apis::SpannerV1::Expr] attr_accessor :condition # Specifies the identities requesting access for a Cloud Platform resource. # `members` can have the following values: # * `allUsers`: A special identifier that represents anyone who is # on the internet; with or without a Google account. # * `allAuthenticatedUsers`: A special identifier that represents anyone # who is authenticated with a Google account or a service account. # * `user:`emailid``: An email address that represents a specific Google # account. For example, `alice@example.com` . # * `serviceAccount:`emailid``: An email address that represents a service # account. For example, `my-other-app@appspot.gserviceaccount.com`. # * `group:`emailid``: An email address that represents a Google group. # For example, `admins@example.com`. # * `deleted:user:`emailid`?uid=`uniqueid``: An email address (plus unique # identifier) representing a user that has been recently deleted. For # example, `alice@example.com?uid=123456789012345678901`. If the user is # recovered, this value reverts to `user:`emailid`` and the recovered user # retains the role in the binding. # * `deleted:serviceAccount:`emailid`?uid=`uniqueid``: An email address (plus # unique identifier) representing a service account that has been recently # deleted. For example, # `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. # If the service account is undeleted, this value reverts to # `serviceAccount:`emailid`` and the undeleted service account retains the # role in the binding. # * `deleted:group:`emailid`?uid=`uniqueid``: An email address (plus unique # identifier) representing a Google group that has been recently # deleted. For example, `admins@example.com?uid=123456789012345678901`. If # the group is recovered, this value reverts to `group:`emailid`` and the # recovered group retains the role in the binding. # * `domain:`domain``: The G Suite domain (primary) that represents all the # users of that domain. For example, `google.com` or `example.com`. # Corresponds to the JSON property `members` # @return [Array] attr_accessor :members # Role that is assigned to `members`. # For example, `roles/viewer`, `roles/editor`, or `roles/owner`. # Corresponds to the JSON property `role` # @return [String] attr_accessor :role def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @condition = args[:condition] if args.key?(:condition) @members = args[:members] if args.key?(:members) @role = args[:role] if args.key?(:role) end end # Metadata associated with a parent-child relationship appearing in a # PlanNode. class ChildLink include Google::Apis::Core::Hashable # The node to which the link points. # Corresponds to the JSON property `childIndex` # @return [Fixnum] attr_accessor :child_index # The type of the link. For example, in Hash Joins this could be used to # distinguish between the build child and the probe child, or in the case # of the child being an output variable, to represent the tag associated # with the output variable. # Corresponds to the JSON property `type` # @return [String] attr_accessor :type # Only present if the child node is SCALAR and corresponds # to an output variable of the parent node. The field carries the name of # the output variable. # For example, a `TableScan` operator that reads rows from a table will # have child links to the `SCALAR` nodes representing the output variables # created for each column that is read by the operator. The corresponding # `variable` fields will be set to the variable names assigned to the # columns. # Corresponds to the JSON property `variable` # @return [String] attr_accessor :variable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @child_index = args[:child_index] if args.key?(:child_index) @type = args[:type] if args.key?(:type) @variable = args[:variable] if args.key?(:variable) end end # The request for Commit. class CommitRequest include Google::Apis::Core::Hashable # The mutations to be executed when this transaction commits. All # mutations are applied atomically, in the order they appear in # this list. # Corresponds to the JSON property `mutations` # @return [Array] attr_accessor :mutations # # Transactions # Each session can have at most one active transaction at a time (note that # standalone reads and queries use a transaction internally and do count # towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. # It is not necessary to create a new session for each transaction. # # Transaction Modes # Cloud Spanner supports three transaction modes: # 1. Locking read-write. This type of transaction is the only way # to write data into Cloud Spanner. These transactions rely on # pessimistic locking and, if necessary, two-phase commit. # Locking read-write transactions may abort, requiring the # application to retry. # 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow # writes. Snapshot read-only transactions can be configured to # read at timestamps in the past. Snapshot read-only # transactions do not need to be committed. # 3. Partitioned DML. This type of transaction is used to execute # a single Partitioned DML statement. Partitioned DML partitions # the key space and runs the DML statement over each partition # in parallel using separate, internal transactions that commit # independently. Partitioned DML transactions do not need to be # committed. # For transactions that only read, snapshot read-only transactions # provide simpler semantics and are almost always faster. In # particular, read-only transactions do not take locks, so they do # not conflict with read-write transactions. As a consequence of not # taking locks, they also do not abort, so retry loops are not needed. # Transactions may only read/write data in a single database. They # may, however, read/write data in different tables within that # database. # ## Locking Read-Write Transactions # Locking transactions may be used to atomically read-modify-write # data anywhere in a database. This type of transaction is externally # consistent. # Clients should attempt to minimize the amount of time a transaction # is active. Faster transactions commit with higher probability # and cause less contention. Cloud Spanner attempts to keep read locks # active as long as the transaction continues to do reads, and the # transaction has not been terminated by # Commit or # Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a # transaction's locks and abort it. # Conceptually, a read-write transaction consists of zero or more # reads or SQL statements followed by # Commit. At any time before # Commit, the client can send a # Rollback request to abort the # transaction. # ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired # are still valid at commit time, and it is able to acquire write # locks for all writes. Cloud Spanner can abort the transaction for any # reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees # that the transaction has not modified any user data in Cloud Spanner. # Unless the transaction commits, Cloud Spanner makes no guarantees about # how long the transaction's locks were held for. It is an error to # use Cloud Spanner locks for any sort of mutual exclusion other than # between Cloud Spanner transactions themselves. # ### Retrying Aborted Transactions # When a transaction aborts, the application can choose to retry the # whole transaction again. To maximize the chances of successfully # committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock # priority increases with each consecutive abort, meaning that each # attempt has a slightly better chance of success than the previous. # Under some circumstances (e.g., many transactions attempting to # modify the same row(s)), a transaction can abort many times in a # short period before successfully committing. Thus, it is not a good # idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent # retrying. # ### Idle Transactions # A transaction is considered idle if it has no outstanding reads or # SQL queries and has not started a read or SQL query within the last 10 # seconds. Idle transactions can be aborted by Cloud Spanner so that they # don't hold on to locks indefinitely. In that case, the commit will # fail with error `ABORTED`. # If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the # transaction from becoming idle. # ## Snapshot Read-Only Transactions # Snapshot read-only transactions provides a simpler method than # locking read-write transactions for doing several consistent # reads. However, this type of transaction does not support writes. # Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that # timestamp. Since they do not acquire locks, they do not block # concurrent read-write transactions. # Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read # timestamp is garbage collected; however, the default garbage # collection policy is generous enough that most applications do not # need to worry about this in practice. # Snapshot read-only transactions do not need to call # Commit or # Rollback (and in fact are not # permitted to do so). # To execute a snapshot transaction, the client specifies a timestamp # bound, which tells Cloud Spanner how to choose a read timestamp. # The types of timestamp bound are: # - Strong (the default). # - Bounded staleness. # - Exact staleness. # If the Cloud Spanner database to be read is geographically distributed, # stale read-only transactions can execute more quickly than strong # or read-write transaction, because they are able to execute far # from the leader replica. # Each type of timestamp bound is discussed in detail below. # ### Strong # Strong reads are guaranteed to see the effects of all transactions # that have committed before the start of the read. Furthermore, all # rows yielded by a single read are consistent with each other -- if # any part of the read observes a transaction, all parts of the read # see the transaction. # Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are # concurrent writes. If consistency across reads is required, the # reads should be executed within a transaction or at an exact read # timestamp. # See TransactionOptions.ReadOnly.strong. # ### Exact Staleness # These timestamp bounds execute reads at a user-specified # timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe # modifications done by all transactions with a commit timestamp <= # the read timestamp, and observe none of the modifications done by # transactions with a larger commit timestamp. They will block until # all conflicting transactions that may be assigned commit timestamps # <= the read timestamp have finished. # The timestamp can either be expressed as an absolute Cloud Spanner commit # timestamp or a staleness relative to the current time. # These modes do not require a "negotiation phase" to pick a # timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, # boundedly stale reads usually return fresher results. # See TransactionOptions.ReadOnly.read_timestamp and # TransactionOptions.ReadOnly.exact_staleness. # ### Bounded Staleness # Bounded staleness modes allow Cloud Spanner to pick the read timestamp, # subject to a user-provided staleness bound. Cloud Spanner chooses the # newest timestamp within the staleness bound that allows execution # of the reads at the closest available replica without blocking. # All rows yielded are consistent with each other -- if any part of # the read observes a transaction, all parts of the read see the # transaction. Boundedly stale reads are not repeatable: two stale # reads, even if they use the same staleness bound, can execute at # different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase # negotiates a timestamp among all replicas needed to serve the # read. In the second phase, reads are executed at the negotiated # timestamp. # As a result of the two phase execution, bounded staleness reads are # usually a little slower than comparable exact staleness # reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. # Because the timestamp negotiation requires up-front knowledge of # which rows will be read, it can only be used with single-use # read-only transactions. # See TransactionOptions.ReadOnly.max_staleness and # TransactionOptions.ReadOnly.min_read_timestamp. # ### Old Read Timestamps and Garbage Collection # Cloud Spanner continuously garbage collects deleted and overwritten data # in the background to reclaim storage space. This process is known # as "version GC". By default, version GC reclaims versions after they # are one hour old. Because of this, Cloud Spanner cannot perform reads # at read timestamps more than one hour in the past. This # restriction also applies to in-progress reads and/or SQL queries whose # timestamp become too old while executing. Reads and SQL queries with # too-old read timestamps fail with the error `FAILED_PRECONDITION`. # ## Partitioned DML Transactions # Partitioned DML transactions are used to execute DML statements with a # different execution strategy that provides different, and often better, # scalability properties for large, table-wide operations than DML in a # ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, # should prefer using ReadWrite transactions. # Partitioned DML partitions the keyspace and runs the DML statement on each # partition in separate, internal transactions. These transactions commit # automatically when complete, and run independently from one another. # To reduce lock contention, this execution strategy only acquires read locks # on rows that match the WHERE clause of the statement. Additionally, the # smaller per-partition transactions hold locks for less time. # That said, Partitioned DML is not a drop-in replacement for standard DML used # in ReadWrite transactions. # - The DML statement must be fully-partitionable. Specifically, the statement # must be expressible as the union of many statements which each access only # a single row of the table. # - The statement is not applied atomically to all rows of the table. Rather, # the statement is applied atomically to partitions of the table, in # independent transactions. Secondary index rows are updated atomically # with the base table rows. # - Partitioned DML does not guarantee exactly-once execution semantics # against a partition. The statement will be applied at least once to each # partition. It is strongly recommended that the DML statement should be # idempotent to avoid unexpected results. For instance, it is potentially # dangerous to run a statement such as # `UPDATE table SET column = column + 1` as it could be run multiple times # against some rows. # - The partitions are committed automatically - there is no support for # Commit or Rollback. If the call returns an error, or if the client issuing # the ExecuteSql call dies, it is possible that some rows had the statement # executed on them successfully. It is also possible that statement was # never executed against other rows. # - Partitioned DML transactions may only contain the execution of a single # DML statement via ExecuteSql or ExecuteStreamingSql. # - If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the # operation is stopped at that point and an error is returned. It is # possible that at this point, some partitions have been committed (or even # committed multiple times), and other partitions have not been run at all. # Given the above, Partitioned DML is good fit for large, database-wide, # operations that are idempotent, such as deleting old rows from a very large # table. # Corresponds to the JSON property `singleUseTransaction` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :single_use_transaction # Commit a previously-started transaction. # Corresponds to the JSON property `transactionId` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :transaction_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @mutations = args[:mutations] if args.key?(:mutations) @single_use_transaction = args[:single_use_transaction] if args.key?(:single_use_transaction) @transaction_id = args[:transaction_id] if args.key?(:transaction_id) end end # The response for Commit. class CommitResponse include Google::Apis::Core::Hashable # The Cloud Spanner timestamp at which the transaction committed. # Corresponds to the JSON property `commitTimestamp` # @return [String] attr_accessor :commit_timestamp def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @commit_timestamp = args[:commit_timestamp] if args.key?(:commit_timestamp) end end # Metadata type for the operation returned by # CreateBackup. class CreateBackupMetadata include Google::Apis::Core::Hashable # The time at which cancellation of this operation was received. # Operations.CancelOperation # starts asynchronous cancellation on a long-running operation. The server # makes a best effort to cancel the operation, but success is not guaranteed. # Clients can use # Operations.GetOperation or # other methods to check whether the cancellation succeeded or whether the # operation completed despite cancellation. On successful cancellation, # the operation is not deleted; instead, it becomes an operation with # an Operation.error value with a google.rpc.Status.code of 1, # corresponding to `Code.CANCELLED`. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The name of the database the backup is created from. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # The name of the backup being created. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Encapsulates progress related information for a Cloud Spanner long # running operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @database = args[:database] if args.key?(:database) @name = args[:name] if args.key?(:name) @progress = args[:progress] if args.key?(:progress) end end # Metadata type for the operation returned by # CreateDatabase. class CreateDatabaseMetadata include Google::Apis::Core::Hashable # The database being created. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @database = args[:database] if args.key?(:database) end end # The request for CreateDatabase. class CreateDatabaseRequest include Google::Apis::Core::Hashable # Required. A `CREATE DATABASE` statement, which specifies the ID of the # new database. The database ID must conform to the regular expression # `a-z*[a-z0-9]` and be between 2 and 30 characters in length. # If the database ID is a reserved word or if it contains a hyphen, the # database ID must be enclosed in backticks (`` ` ``). # Corresponds to the JSON property `createStatement` # @return [String] attr_accessor :create_statement # Optional. A list of DDL statements to run inside the newly created # database. Statements can create tables, indexes, etc. These # statements execute atomically with the creation of the database: # if there is an error in any statement, the database is not created. # Corresponds to the JSON property `extraStatements` # @return [Array] attr_accessor :extra_statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_statement = args[:create_statement] if args.key?(:create_statement) @extra_statements = args[:extra_statements] if args.key?(:extra_statements) end end # Metadata type for the operation returned by # CreateInstance. class CreateInstanceMetadata include Google::Apis::Core::Hashable # The time at which this operation was cancelled. If set, this operation is # in the process of undoing itself (which is guaranteed to succeed) and # cannot be cancelled again. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The time at which this operation failed or was completed successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # The time at which the # CreateInstance request was # received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @end_time = args[:end_time] if args.key?(:end_time) @instance = args[:instance] if args.key?(:instance) @start_time = args[:start_time] if args.key?(:start_time) end end # The request for CreateInstance. class CreateInstanceRequest include Google::Apis::Core::Hashable # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # Required. The ID of the instance to create. Valid identifiers are of the # form `a-z*[a-z0-9]` and must be between 2 and 64 characters in # length. # Corresponds to the JSON property `instanceId` # @return [String] attr_accessor :instance_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance = args[:instance] if args.key?(:instance) @instance_id = args[:instance_id] if args.key?(:instance_id) end end # The request for CreateSession. class CreateSessionRequest include Google::Apis::Core::Hashable # A session in the Cloud Spanner API. # Corresponds to the JSON property `session` # @return [Google::Apis::SpannerV1::Session] attr_accessor :session def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session = args[:session] if args.key?(:session) end end # A Cloud Spanner database. class Database include Google::Apis::Core::Hashable # Output only. If exists, the time at which the database creation started. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Required. The name of the database. Values are of the form # `projects//instances//databases/`, # where `` is as specified in the `CREATE DATABASE` # statement. This name can be passed to other API methods to # identify the database. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Information about the database restore. # Corresponds to the JSON property `restoreInfo` # @return [Google::Apis::SpannerV1::RestoreInfo] attr_accessor :restore_info # Output only. The current database state. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_time = args[:create_time] if args.key?(:create_time) @name = args[:name] if args.key?(:name) @restore_info = args[:restore_info] if args.key?(:restore_info) @state = args[:state] if args.key?(:state) end end # Arguments to delete operations. class Delete include Google::Apis::Core::Hashable # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All # the keys are expected to be in the same table or index. The keys need # not be sorted in any particular way. # If the same key is specified multiple times in the set (for example # if two ranges, two keys, or a key and a range overlap), Cloud Spanner # behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # Required. The table whose rows will be deleted. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @key_set = args[:key_set] if args.key?(:key_set) @table = args[:table] if args.key?(:table) end end # A generic empty message that you can re-use to avoid defining duplicated # empty messages in your APIs. A typical example is to use it as the request # or the response type of an API method. For instance: # service Foo ` # rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); # ` # The JSON representation for `Empty` is empty JSON object ````. class Empty include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # The request for ExecuteBatchDml. class ExecuteBatchDmlRequest include Google::Apis::Core::Hashable # Required. A per-transaction sequence number used to identify this request. # This field # makes each request idempotent such that if the request is received multiple # times, at most one will succeed. # The sequence number must be monotonically increasing within the # transaction. If a request arrives for the first time with an out-of-order # sequence number, the transaction may be aborted. Replays of previously # handled requests will yield the same response as the first execution. # Corresponds to the JSON property `seqno` # @return [Fixnum] attr_accessor :seqno # Required. The list of statements to execute in this batch. Statements are # executed # serially, such that the effects of statement `i` are visible to statement # `i+1`. Each statement must be a DML statement. Execution stops at the # first failed statement; the remaining statements are not executed. # Callers must provide at least one statement. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @seqno = args[:seqno] if args.key?(:seqno) @statements = args[:statements] if args.key?(:statements) @transaction = args[:transaction] if args.key?(:transaction) end end # The response for ExecuteBatchDml. Contains a list # of ResultSet messages, one for each DML statement that has successfully # executed, in the same order as the statements in the request. If a statement # fails, the status in the response body identifies the cause of the failure. # To check for DML statements that failed, use the following approach: # 1. Check the status in the response message. The google.rpc.Code enum # value `OK` indicates that all statements were executed successfully. # 2. If the status was not `OK`, check the number of result sets in the # response. If the response contains `N` ResultSet messages, then # statement `N+1` in the request failed. # Example 1: # * Request: 5 DML statements, all executed successfully. # * Response: 5 ResultSet messages, with the status `OK`. # Example 2: # * Request: 5 DML statements. The third statement has a syntax error. # * Response: 2 ResultSet messages, and a syntax error (`INVALID_ARGUMENT`) # status. The number of ResultSet messages indicates that the third # statement failed, and the fourth and fifth statements were not executed. class ExecuteBatchDmlResponse include Google::Apis::Core::Hashable # One ResultSet for each statement in the request that ran successfully, # in the same order as the statements in the request. Each ResultSet does # not contain any rows. The ResultSetStats in each ResultSet contain # the number of rows modified by the statement. # Only the first ResultSet in the response contains valid # ResultSetMetadata. # Corresponds to the JSON property `resultSets` # @return [Array] attr_accessor :result_sets # The `Status` type defines a logical error model that is suitable for # different programming environments, including REST APIs and RPC APIs. It is # used by [gRPC](https://github.com/grpc). Each `Status` message contains # three pieces of data: error code, error message, and error details. # You can find out more about this error model and how to work with it in the # [API Design Guide](https://cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::SpannerV1::Status] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @result_sets = args[:result_sets] if args.key?(:result_sets) @status = args[:status] if args.key?(:status) end end # The request for ExecuteSql and # ExecuteStreamingSql. class ExecuteSqlRequest include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type # from a JSON value. For example, values of type `BYTES` and values # of type `STRING` both appear in params as JSON strings. # In these cases, `param_types` can be used to specify the exact # SQL type for some or all of the SQL statement parameters. See the # definition of Type for more information # about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the SQL string. # A parameter placeholder consists of the `@` character followed by the # parameter name (for example, `@firstName`). Parameter names must conform # to the naming requirements of identifiers as specified at # https://cloud.google.com/spanner/docs/lexical#identifiers. # Parameters can appear anywhere that a literal value is expected. The same # parameter name can be used more than once, for example: # `"WHERE id > @msg_id AND id < @msg_id + 100"` # It is an error to execute a SQL statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # If present, results will be restricted to the specified partition # previously created using PartitionQuery(). There must be an exact # match for the values of fields common to this message and the # PartitionQueryRequest message used to create this partition_token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token # Used to control the amount of debugging information returned in # ResultSetStats. If partition_token is set, query_mode can only # be set to QueryMode.NORMAL. # Corresponds to the JSON property `queryMode` # @return [String] attr_accessor :query_mode # Query optimizer configuration. # Corresponds to the JSON property `queryOptions` # @return [Google::Apis::SpannerV1::QueryOptions] attr_accessor :query_options # If this request is resuming a previously interrupted SQL statement # execution, `resume_token` should be copied from the last # PartialResultSet yielded before the interruption. Doing this # enables the new SQL statement execution to resume where the last one left # off. The rest of the request parameters must exactly match the # request that yielded this token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # A per-transaction sequence number used to identify this request. This field # makes each request idempotent such that if the request is received multiple # times, at most one will succeed. # The sequence number must be monotonically increasing within the # transaction. If a request arrives for the first time with an out-of-order # sequence number, the transaction may be aborted. Replays of previously # handled requests will yield the same response as the first execution. # Required for DML statements. Ignored for queries. # Corresponds to the JSON property `seqno` # @return [Fixnum] attr_accessor :seqno # Required. The SQL string. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @partition_token = args[:partition_token] if args.key?(:partition_token) @query_mode = args[:query_mode] if args.key?(:query_mode) @query_options = args[:query_options] if args.key?(:query_options) @resume_token = args[:resume_token] if args.key?(:resume_token) @seqno = args[:seqno] if args.key?(:seqno) @sql = args[:sql] if args.key?(:sql) @transaction = args[:transaction] if args.key?(:transaction) end end # Represents a textual expression in the Common Expression Language (CEL) # syntax. CEL is a C-like expression language. The syntax and semantics of CEL # are documented at https://github.com/google/cel-spec. # Example (Comparison): # title: "Summary size limit" # description: "Determines if a summary is less than 100 chars" # expression: "document.summary.size() < 100" # Example (Equality): # title: "Requestor is owner" # description: "Determines if requestor is the document owner" # expression: "document.owner == request.auth.claims.email" # Example (Logic): # title: "Public documents" # description: "Determine whether the document should be publicly visible" # expression: "document.type != 'private' && document.type != 'internal'" # Example (Data Manipulation): # title: "Notification string" # description: "Create a notification string with a timestamp." # expression: "'New message received at ' + string(document.create_time)" # The exact variables and functions that may be referenced within an expression # are determined by the service that evaluates it. See the service # documentation for additional information. class Expr include Google::Apis::Core::Hashable # Optional. Description of the expression. This is a longer text which # describes the expression, e.g. when hovered over it in a UI. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description # Textual representation of an expression in Common Expression Language # syntax. # Corresponds to the JSON property `expression` # @return [String] attr_accessor :expression # Optional. String indicating the location of the expression for error # reporting, e.g. a file name and a position in the file. # Corresponds to the JSON property `location` # @return [String] attr_accessor :location # Optional. Title for the expression, i.e. a short string describing # its purpose. This can be used e.g. in UIs which allow to enter the # expression. # Corresponds to the JSON property `title` # @return [String] attr_accessor :title def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @description = args[:description] if args.key?(:description) @expression = args[:expression] if args.key?(:expression) @location = args[:location] if args.key?(:location) @title = args[:title] if args.key?(:title) end end # Message representing a single field of a struct. class Field include Google::Apis::Core::Hashable # The name of the field. For reads, this is the column name. For # SQL queries, it is the column alias (e.g., `"Word"` in the # query `"SELECT 'hello' AS Word"`), or the column name (e.g., # `"ColName"` in the query `"SELECT ColName FROM Table"`). Some # columns might have an empty name (e.g., !"SELECT # UPPER(ColName)"`). Note that a query result can contain # multiple fields with the same name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. # Corresponds to the JSON property `type` # @return [Google::Apis::SpannerV1::Type] attr_accessor :type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @type = args[:type] if args.key?(:type) end end # The response for GetDatabaseDdl. class GetDatabaseDdlResponse include Google::Apis::Core::Hashable # A list of formatted DDL statements defining the schema of the database # specified in the request. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @statements = args[:statements] if args.key?(:statements) end end # Request message for `GetIamPolicy` method. class GetIamPolicyRequest include Google::Apis::Core::Hashable # Encapsulates settings provided to GetIamPolicy. # Corresponds to the JSON property `options` # @return [Google::Apis::SpannerV1::GetPolicyOptions] attr_accessor :options def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @options = args[:options] if args.key?(:options) end end # Encapsulates settings provided to GetIamPolicy. class GetPolicyOptions include Google::Apis::Core::Hashable # Optional. The policy format version to be returned. # Valid values are 0, 1, and 3. Requests specifying an invalid value will be # rejected. # Requests for policies with any conditional bindings must specify version 3. # Policies without any conditional bindings may specify any valid value or # leave the field unset. # To learn which resources support conditions in their IAM policies, see the # [IAM # documentation](https://cloud.google.com/iam/help/conditions/resource-policies). # Corresponds to the JSON property `requestedPolicyVersion` # @return [Fixnum] attr_accessor :requested_policy_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @requested_policy_version = args[:requested_policy_version] if args.key?(:requested_policy_version) end end # An isolated set of Cloud Spanner resources on which databases can be hosted. class Instance include Google::Apis::Core::Hashable # Required. The name of the instance's configuration. Values are of the form # `projects//instanceConfigs/`. See # also InstanceConfig and # ListInstanceConfigs. # Corresponds to the JSON property `config` # @return [String] attr_accessor :config # Required. The descriptive name for this instance as it appears in UIs. # Must be unique per project and between 4 and 30 characters in length. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # Deprecated. This field is not populated. # Corresponds to the JSON property `endpointUris` # @return [Array] attr_accessor :endpoint_uris # Cloud Labels are a flexible and lightweight mechanism for organizing cloud # resources into groups that reflect a customer's organizational needs and # deployment strategies. Cloud Labels can be used to filter collections of # resources. They can be used to control how resource metrics are aggregated. # And they can be used as arguments to policy management rules (e.g. route, # firewall, load balancing, etc.). # * Label keys must be between 1 and 63 characters long and must conform to # the following regular expression: `[a-z]([-a-z0-9]*[a-z0-9])?`. # * Label values must be between 0 and 63 characters long and must conform # to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. # * No more than 64 labels can be associated with a given resource. # See https://goo.gl/xmQnxf for more information on and examples of labels. # If you plan to use labels in your own code, please note that additional # characters may be allowed in the future. And so you are advised to use an # internal label representation, such as JSON, which doesn't rely upon # specific characters being disallowed. For example, representing labels # as the string: name + "_" + value would prove problematic if we were to # allow "_" in a future release. # Corresponds to the JSON property `labels` # @return [Hash] attr_accessor :labels # Required. A unique identifier for the instance, which cannot be changed # after the instance is created. Values are of the form # `projects//instances/a-z*[a-z0-9]`. The final # segment of the name must be between 2 and 64 characters in length. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The number of nodes allocated to this instance. This # may be zero in API responses for instances that are not yet in state # `READY`. # See [the # documentation](https://cloud.google.com/spanner/docs/instances#node_count) # for more information about nodes. # Corresponds to the JSON property `nodeCount` # @return [Fixnum] attr_accessor :node_count # Output only. The current instance state. For # CreateInstance, the state must be # either omitted or set to `CREATING`. For # UpdateInstance, the state must be # either omitted or set to `READY`. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @config = args[:config] if args.key?(:config) @display_name = args[:display_name] if args.key?(:display_name) @endpoint_uris = args[:endpoint_uris] if args.key?(:endpoint_uris) @labels = args[:labels] if args.key?(:labels) @name = args[:name] if args.key?(:name) @node_count = args[:node_count] if args.key?(:node_count) @state = args[:state] if args.key?(:state) end end # A possible configuration for a Cloud Spanner instance. Configurations # define the geographic placement of nodes and their replication. class InstanceConfig include Google::Apis::Core::Hashable # The name of this instance configuration as it appears in UIs. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # A unique identifier for the instance configuration. Values # are of the form # `projects//instanceConfigs/a-z*` # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The geographic placement of nodes in this instance configuration and their # replication properties. # Corresponds to the JSON property `replicas` # @return [Array] attr_accessor :replicas def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @display_name = args[:display_name] if args.key?(:display_name) @name = args[:name] if args.key?(:name) @replicas = args[:replicas] if args.key?(:replicas) end end # KeyRange represents a range of rows in a table or index. # A range has a start key and an end key. These keys can be open or # closed, indicating if the range includes rows with that key. # Keys are represented by lists, where the ith value in the list # corresponds to the ith component of the table or index primary key. # Individual values are encoded as described # here. # For example, consider the following table definition: # CREATE TABLE UserEvents ( # UserName STRING(MAX), # EventDate STRING(10) # ) PRIMARY KEY(UserName, EventDate); # The following keys name rows in this table: # "Bob", "2014-09-23" # Since the `UserEvents` table's `PRIMARY KEY` clause names two # columns, each `UserEvents` key has two elements; the first is the # `UserName`, and the second is the `EventDate`. # Key ranges with multiple components are interpreted # lexicographically by component using the table or index key's declared # sort order. For example, the following range returns all events for # user `"Bob"` that occurred in the year 2015: # "start_closed": ["Bob", "2015-01-01"] # "end_closed": ["Bob", "2015-12-31"] # Start and end keys can omit trailing key components. This affects the # inclusion and exclusion of rows that exactly match the provided key # components: if the key is closed, then rows that exactly match the # provided components are included; if the key is open, then rows # that exactly match are not included. # For example, the following range includes all events for `"Bob"` that # occurred during and after the year 2000: # "start_closed": ["Bob", "2000-01-01"] # "end_closed": ["Bob"] # The next example retrieves all events for `"Bob"`: # "start_closed": ["Bob"] # "end_closed": ["Bob"] # To retrieve events before the year 2000: # "start_closed": ["Bob"] # "end_open": ["Bob", "2000-01-01"] # The following range includes all rows in the table: # "start_closed": [] # "end_closed": [] # This range returns all users whose `UserName` begins with any # character from A to C: # "start_closed": ["A"] # "end_open": ["D"] # This range returns all users whose `UserName` begins with B: # "start_closed": ["B"] # "end_open": ["C"] # Key ranges honor column sort order. For example, suppose a table is # defined as follows: # CREATE TABLE DescendingSortedTable ` # Key INT64, # ... # ) PRIMARY KEY(Key DESC); # The following range retrieves all rows with key values between 1 # and 100 inclusive: # "start_closed": ["100"] # "end_closed": ["1"] # Note that 100 is passed as the start, and 1 is passed as the end, # because `Key` is a descending column in the schema. class KeyRange include Google::Apis::Core::Hashable # If the end is closed, then the range includes all rows whose # first `len(end_closed)` key columns exactly match `end_closed`. # Corresponds to the JSON property `endClosed` # @return [Array] attr_accessor :end_closed # If the end is open, then the range excludes rows whose first # `len(end_open)` key columns exactly match `end_open`. # Corresponds to the JSON property `endOpen` # @return [Array] attr_accessor :end_open # If the start is closed, then the range includes all rows whose # first `len(start_closed)` key columns exactly match `start_closed`. # Corresponds to the JSON property `startClosed` # @return [Array] attr_accessor :start_closed # If the start is open, then the range excludes rows whose first # `len(start_open)` key columns exactly match `start_open`. # Corresponds to the JSON property `startOpen` # @return [Array] attr_accessor :start_open def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @end_closed = args[:end_closed] if args.key?(:end_closed) @end_open = args[:end_open] if args.key?(:end_open) @start_closed = args[:start_closed] if args.key?(:start_closed) @start_open = args[:start_open] if args.key?(:start_open) end end # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All # the keys are expected to be in the same table or index. The keys need # not be sorted in any particular way. # If the same key is specified multiple times in the set (for example # if two ranges, two keys, or a key and a range overlap), Cloud Spanner # behaves as if the key were only specified once. class KeySet include Google::Apis::Core::Hashable # For convenience `all` can be set to `true` to indicate that this # `KeySet` matches all keys in the table or index. Note that any keys # specified in `keys` or `ranges` are only yielded once. # Corresponds to the JSON property `all` # @return [Boolean] attr_accessor :all alias_method :all?, :all # A list of specific keys. Entries in `keys` should have exactly as # many elements as there are columns in the primary or index key # with which this `KeySet` is used. Individual key values are # encoded as described here. # Corresponds to the JSON property `keys` # @return [Array>] attr_accessor :keys # A list of key ranges. See KeyRange for more information about # key range specifications. # Corresponds to the JSON property `ranges` # @return [Array] attr_accessor :ranges def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @all = args[:all] if args.key?(:all) @keys = args[:keys] if args.key?(:keys) @ranges = args[:ranges] if args.key?(:ranges) end end # The response for # ListBackupOperations. class ListBackupOperationsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent # ListBackupOperations # call to fetch more of the matching metadata. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of matching backup long-running # operations. Each operation's name will be # prefixed by the backup's name and the operation's # metadata will be of type # CreateBackupMetadata. Operations returned include those that are # pending or have completed/failed/canceled within the last 7 days. # Operations returned are ordered by # `operation.metadata.value.progress.start_time` in descending order starting # from the most recently started operation. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListBackups. class ListBackupsResponse include Google::Apis::Core::Hashable # The list of matching backups. Backups returned are ordered by `create_time` # in descending order, starting from the most recent `create_time`. # Corresponds to the JSON property `backups` # @return [Array] attr_accessor :backups # `next_page_token` can be sent in a subsequent # ListBackups call to fetch more # of the matching backups. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backups = args[:backups] if args.key?(:backups) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for # ListDatabaseOperations. class ListDatabaseOperationsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent # ListDatabaseOperations # call to fetch more of the matching metadata. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of matching database long-running # operations. Each operation's name will be # prefixed by the database's name. The operation's # metadata field type # `metadata.type_url` describes the type of the metadata. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListDatabases. class ListDatabasesResponse include Google::Apis::Core::Hashable # Databases that matched the request. # Corresponds to the JSON property `databases` # @return [Array] attr_accessor :databases # `next_page_token` can be sent in a subsequent # ListDatabases call to fetch more # of the matching databases. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @databases = args[:databases] if args.key?(:databases) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for ListInstanceConfigs. class ListInstanceConfigsResponse include Google::Apis::Core::Hashable # The list of requested instance configurations. # Corresponds to the JSON property `instanceConfigs` # @return [Array] attr_accessor :instance_configs # `next_page_token` can be sent in a subsequent # ListInstanceConfigs call to # fetch more of the matching instance configurations. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance_configs = args[:instance_configs] if args.key?(:instance_configs) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for ListInstances. class ListInstancesResponse include Google::Apis::Core::Hashable # The list of requested instances. # Corresponds to the JSON property `instances` # @return [Array] attr_accessor :instances # `next_page_token` can be sent in a subsequent # ListInstances call to fetch more # of the matching instances. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instances = args[:instances] if args.key?(:instances) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response message for Operations.ListOperations. class ListOperationsResponse include Google::Apis::Core::Hashable # The standard List next-page token. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # A list of operations that matches the specified filter in the request. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListSessions. class ListSessionsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent # ListSessions call to fetch more of the matching # sessions. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of requested sessions. # Corresponds to the JSON property `sessions` # @return [Array] attr_accessor :sessions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @sessions = args[:sessions] if args.key?(:sessions) end end # A modification to one or more Cloud Spanner rows. Mutations can be # applied to a Cloud Spanner database by sending them in a # Commit call. class Mutation include Google::Apis::Core::Hashable # Arguments to delete operations. # Corresponds to the JSON property `delete` # @return [Google::Apis::SpannerV1::Delete] attr_accessor :delete # Arguments to insert, update, insert_or_update, and # replace operations. # Corresponds to the JSON property `insert` # @return [Google::Apis::SpannerV1::Write] attr_accessor :insert # Arguments to insert, update, insert_or_update, and # replace operations. # Corresponds to the JSON property `insertOrUpdate` # @return [Google::Apis::SpannerV1::Write] attr_accessor :insert_or_update # Arguments to insert, update, insert_or_update, and # replace operations. # Corresponds to the JSON property `replace` # @return [Google::Apis::SpannerV1::Write] attr_accessor :replace # Arguments to insert, update, insert_or_update, and # replace operations. # Corresponds to the JSON property `update` # @return [Google::Apis::SpannerV1::Write] attr_accessor :update def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @delete = args[:delete] if args.key?(:delete) @insert = args[:insert] if args.key?(:insert) @insert_or_update = args[:insert_or_update] if args.key?(:insert_or_update) @replace = args[:replace] if args.key?(:replace) @update = args[:update] if args.key?(:update) end end # This resource represents a long-running operation that is the result of a # network API call. class Operation include Google::Apis::Core::Hashable # If the value is `false`, it means the operation is still in progress. # If `true`, the operation is completed, and either `error` or `response` is # available. # Corresponds to the JSON property `done` # @return [Boolean] attr_accessor :done alias_method :done?, :done # The `Status` type defines a logical error model that is suitable for # different programming environments, including REST APIs and RPC APIs. It is # used by [gRPC](https://github.com/grpc). Each `Status` message contains # three pieces of data: error code, error message, and error details. # You can find out more about this error model and how to work with it in the # [API Design Guide](https://cloud.google.com/apis/design/errors). # Corresponds to the JSON property `error` # @return [Google::Apis::SpannerV1::Status] attr_accessor :error # Service-specific metadata associated with the operation. It typically # contains progress information and common metadata such as create time. # Some services might not provide such metadata. Any method that returns a # long-running operation should document the metadata type, if any. # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # The server-assigned name, which is only unique within the same service that # originally returns it. If you use the default HTTP mapping, the # `name` should be a resource name ending with `operations/`unique_id``. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The normal response of the operation in case of success. If the original # method returns no data on success, such as `Delete`, the response is # `google.protobuf.Empty`. If the original method is standard # `Get`/`Create`/`Update`, the response should be the resource. For other # methods, the response should have the type `XxxResponse`, where `Xxx` # is the original method name. For example, if the original method name # is `TakeSnapshot()`, the inferred response type is # `TakeSnapshotResponse`. # Corresponds to the JSON property `response` # @return [Hash] attr_accessor :response def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @done = args[:done] if args.key?(:done) @error = args[:error] if args.key?(:error) @metadata = args[:metadata] if args.key?(:metadata) @name = args[:name] if args.key?(:name) @response = args[:response] if args.key?(:response) end end # Encapsulates progress related information for a Cloud Spanner long # running operation. class OperationProgress include Google::Apis::Core::Hashable # If set, the time at which this operation failed or was completed # successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # Percent completion of the operation. # Values are between 0 and 100 inclusive. # Corresponds to the JSON property `progressPercent` # @return [Fixnum] attr_accessor :progress_percent # Time the request was received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @end_time = args[:end_time] if args.key?(:end_time) @progress_percent = args[:progress_percent] if args.key?(:progress_percent) @start_time = args[:start_time] if args.key?(:start_time) end end # Metadata type for the long-running operation used to track the progress # of optimizations performed on a newly restored database. This long-running # operation is automatically created by the system after the successful # completion of a database restore, and cannot be cancelled. class OptimizeRestoredDatabaseMetadata include Google::Apis::Core::Hashable # Name of the restored database being optimized. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Encapsulates progress related information for a Cloud Spanner long # running operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @progress = args[:progress] if args.key?(:progress) end end # Partial results from a streaming read or SQL query. Streaming reads and # SQL queries better tolerate large result sets, large rows, and large # values, but are a little trickier to consume. class PartialResultSet include Google::Apis::Core::Hashable # If true, then the final value in values is chunked, and must # be combined with more values from subsequent `PartialResultSet`s # to obtain a complete field value. # Corresponds to the JSON property `chunkedValue` # @return [Boolean] attr_accessor :chunked_value alias_method :chunked_value?, :chunked_value # Metadata about a ResultSet or PartialResultSet. # Corresponds to the JSON property `metadata` # @return [Google::Apis::SpannerV1::ResultSetMetadata] attr_accessor :metadata # Streaming calls might be interrupted for a variety of reasons, such # as TCP connection loss. If this occurs, the stream of results can # be resumed by re-sending the original request and including # `resume_token`. Note that executing any other transaction in the # same session invalidates the token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # Additional statistics about a ResultSet or PartialResultSet. # Corresponds to the JSON property `stats` # @return [Google::Apis::SpannerV1::ResultSetStats] attr_accessor :stats # A streamed result set consists of a stream of values, which might # be split into many `PartialResultSet` messages to accommodate # large rows and/or large values. Every N complete values defines a # row, where N is equal to the number of entries in # metadata.row_type.fields. # Most values are encoded based on type as described # here. # It is possible that the last value in values is "chunked", # meaning that the rest of the value is sent in subsequent # `PartialResultSet`(s). This is denoted by the chunked_value # field. Two or more chunked values can be merged to form a # complete value as follows: # * `bool/number/null`: cannot be chunked # * `string`: concatenate the strings # * `list`: concatenate the lists. If the last element in a list is a # `string`, `list`, or `object`, merge it with the first element in # the next list by applying these rules recursively. # * `object`: concatenate the (field name, field value) pairs. If a # field name is duplicated, then apply these rules recursively # to merge the field values. # Some examples of merging: # # Strings are concatenated. # "foo", "bar" => "foobar" # # Lists of non-strings are concatenated. # [2, 3], [4] => [2, 3, 4] # # Lists are concatenated, but the last and first elements are merged # # because they are strings. # ["a", "b"], ["c", "d"] => ["a", "bc", "d"] # # Lists are concatenated, but the last and first elements are merged # # because they are lists. Recursively, the last and first elements # # of the inner lists are merged because they are strings. # ["a", ["b", "c"]], [["d"], "e"] => ["a", ["b", "cd"], "e"] # # Non-overlapping object fields are combined. # `"a": "1"`, `"b": "2"` => `"a": "1", "b": 2"` # # Overlapping object fields are merged. # `"a": "1"`, `"a": "2"` => `"a": "12"` # # Examples of merging objects containing lists of strings. # `"a": ["1"]`, `"a": ["2"]` => `"a": ["12"]` # For a more complete example, suppose a streaming SQL query is # yielding a result set whose rows contain a single string # field. The following `PartialResultSet`s might be yielded: # ` # "metadata": ` ... ` # "values": ["Hello", "W"] # "chunked_value": true # "resume_token": "Af65..." # ` # ` # "values": ["orl"] # "chunked_value": true # "resume_token": "Bqp2..." # ` # ` # "values": ["d"] # "resume_token": "Zx1B..." # ` # This sequence of `PartialResultSet`s encodes two rows, one # containing the field value `"Hello"`, and a second containing the # field value `"World" = "W" + "orl" + "d"`. # Corresponds to the JSON property `values` # @return [Array] attr_accessor :values def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @chunked_value = args[:chunked_value] if args.key?(:chunked_value) @metadata = args[:metadata] if args.key?(:metadata) @resume_token = args[:resume_token] if args.key?(:resume_token) @stats = args[:stats] if args.key?(:stats) @values = args[:values] if args.key?(:values) end end # Information returned for each partition returned in a # PartitionResponse. class Partition include Google::Apis::Core::Hashable # This token can be passed to Read, StreamingRead, ExecuteSql, or # ExecuteStreamingSql requests to restrict the results to those identified by # this partition token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partition_token = args[:partition_token] if args.key?(:partition_token) end end # Options for a PartitionQueryRequest and # PartitionReadRequest. class PartitionOptions include Google::Apis::Core::Hashable # **Note:** This hint is currently ignored by PartitionQuery and # PartitionRead requests. # The desired maximum number of partitions to return. For example, this may # be set to the number of workers available. The default for this option # is currently 10,000. The maximum value is currently 200,000. This is only # a hint. The actual number of partitions returned may be smaller or larger # than this maximum count request. # Corresponds to the JSON property `maxPartitions` # @return [Fixnum] attr_accessor :max_partitions # **Note:** This hint is currently ignored by PartitionQuery and # PartitionRead requests. # The desired data size for each partition generated. The default for this # option is currently 1 GiB. This is only a hint. The actual size of each # partition may be smaller or larger than this size request. # Corresponds to the JSON property `partitionSizeBytes` # @return [Fixnum] attr_accessor :partition_size_bytes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @max_partitions = args[:max_partitions] if args.key?(:max_partitions) @partition_size_bytes = args[:partition_size_bytes] if args.key?(:partition_size_bytes) end end # The request for PartitionQuery class PartitionQueryRequest include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type # from a JSON value. For example, values of type `BYTES` and values # of type `STRING` both appear in params as JSON strings. # In these cases, `param_types` can be used to specify the exact # SQL type for some or all of the SQL query parameters. See the # definition of Type for more information # about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the SQL string. # A parameter placeholder consists of the `@` character followed by the # parameter name (for example, `@firstName`). Parameter names can contain # letters, numbers, and underscores. # Parameters can appear anywhere that a literal value is expected. The same # parameter name can be used more than once, for example: # `"WHERE id > @msg_id AND id < @msg_id + 100"` # It is an error to execute a SQL statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # Options for a PartitionQueryRequest and # PartitionReadRequest. # Corresponds to the JSON property `partitionOptions` # @return [Google::Apis::SpannerV1::PartitionOptions] attr_accessor :partition_options # Required. The query request to generate partitions for. The request will fail # if # the query is not root partitionable. The query plan of a root # partitionable query has a single distributed union operator. A distributed # union operator conceptually divides one or more tables into multiple # splits, remotely evaluates a subquery independently on each split, and # then unions all results. # This must not contain DML commands, such as INSERT, UPDATE, or # DELETE. Use ExecuteStreamingSql with a # PartitionedDml transaction for large, partition-friendly DML operations. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @partition_options = args[:partition_options] if args.key?(:partition_options) @sql = args[:sql] if args.key?(:sql) @transaction = args[:transaction] if args.key?(:transaction) end end # The request for PartitionRead class PartitionReadRequest include Google::Apis::Core::Hashable # The columns of table to be returned for each row matching # this request. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # If non-empty, the name of an index on table. This index is # used instead of the table primary key when interpreting key_set # and sorting result rows. See key_set for further information. # Corresponds to the JSON property `index` # @return [String] attr_accessor :index # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All # the keys are expected to be in the same table or index. The keys need # not be sorted in any particular way. # If the same key is specified multiple times in the set (for example # if two ranges, two keys, or a key and a range overlap), Cloud Spanner # behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # Options for a PartitionQueryRequest and # PartitionReadRequest. # Corresponds to the JSON property `partitionOptions` # @return [Google::Apis::SpannerV1::PartitionOptions] attr_accessor :partition_options # Required. The name of the table in the database to be read. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @index = args[:index] if args.key?(:index) @key_set = args[:key_set] if args.key?(:key_set) @partition_options = args[:partition_options] if args.key?(:partition_options) @table = args[:table] if args.key?(:table) @transaction = args[:transaction] if args.key?(:transaction) end end # The response for PartitionQuery # or PartitionRead class PartitionResponse include Google::Apis::Core::Hashable # Partitions created by this request. # Corresponds to the JSON property `partitions` # @return [Array] attr_accessor :partitions # A transaction. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::Transaction] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partitions = args[:partitions] if args.key?(:partitions) @transaction = args[:transaction] if args.key?(:transaction) end end # Message type to initiate a Partitioned DML transaction. class PartitionedDml include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # Node information for nodes appearing in a QueryPlan.plan_nodes. class PlanNode include Google::Apis::Core::Hashable # List of child node `index`es and their relationship to this parent. # Corresponds to the JSON property `childLinks` # @return [Array] attr_accessor :child_links # The display name for the node. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # The execution statistics associated with the node, contained in a group of # key-value pairs. Only present if the plan was returned as a result of a # profile query. For example, number of executions, number of rows/time per # execution etc. # Corresponds to the JSON property `executionStats` # @return [Hash] attr_accessor :execution_stats # The `PlanNode`'s index in node list. # Corresponds to the JSON property `index` # @return [Fixnum] attr_accessor :index # Used to determine the type of node. May be needed for visualizing # different kinds of nodes differently. For example, If the node is a # SCALAR node, it will have a condensed representation # which can be used to directly embed a description of the node in its # parent. # Corresponds to the JSON property `kind` # @return [String] attr_accessor :kind # Attributes relevant to the node contained in a group of key-value pairs. # For example, a Parameter Reference node could have the following # information in its metadata: # ` # "parameter_reference": "param1", # "parameter_type": "array" # ` # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # Condensed representation of a node and its subtree. Only present for # `SCALAR` PlanNode(s). # Corresponds to the JSON property `shortRepresentation` # @return [Google::Apis::SpannerV1::ShortRepresentation] attr_accessor :short_representation def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @child_links = args[:child_links] if args.key?(:child_links) @display_name = args[:display_name] if args.key?(:display_name) @execution_stats = args[:execution_stats] if args.key?(:execution_stats) @index = args[:index] if args.key?(:index) @kind = args[:kind] if args.key?(:kind) @metadata = args[:metadata] if args.key?(:metadata) @short_representation = args[:short_representation] if args.key?(:short_representation) end end # An Identity and Access Management (IAM) policy, which specifies access # controls for Google Cloud resources. # A `Policy` is a collection of `bindings`. A `binding` binds one or more # `members` to a single `role`. Members can be user accounts, service accounts, # Google groups, and domains (such as G Suite). A `role` is a named list of # permissions; each `role` can be an IAM predefined role or a user-created # custom role. # For some types of Google Cloud resources, a `binding` can also specify a # `condition`, which is a logical expression that allows access to a resource # only if the expression evaluates to `true`. A condition can add constraints # based on attributes of the request, the resource, or both. To learn which # resources support conditions in their IAM policies, see the # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource- # policies). # **JSON example:** # ` # "bindings": [ # ` # "role": "roles/resourcemanager.organizationAdmin", # "members": [ # "user:mike@example.com", # "group:admins@example.com", # "domain:google.com", # "serviceAccount:my-project-id@appspot.gserviceaccount.com" # ] # `, # ` # "role": "roles/resourcemanager.organizationViewer", # "members": [ # "user:eve@example.com" # ], # "condition": ` # "title": "expirable access", # "description": "Does not grant access after Sep 2020", # "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z') # ", # ` # ` # ], # "etag": "BwWWja0YfJA=", # "version": 3 # ` # **YAML example:** # bindings: # - members: # - user:mike@example.com # - group:admins@example.com # - domain:google.com # - serviceAccount:my-project-id@appspot.gserviceaccount.com # role: roles/resourcemanager.organizationAdmin # - members: # - user:eve@example.com # role: roles/resourcemanager.organizationViewer # condition: # title: expirable access # description: Does not grant access after Sep 2020 # expression: request.time < timestamp('2020-10-01T00:00:00.000Z') # - etag: BwWWja0YfJA= # - version: 3 # For a description of IAM and its features, see the # [IAM documentation](https://cloud.google.com/iam/docs/). class Policy include Google::Apis::Core::Hashable # Associates a list of `members` to a `role`. Optionally, may specify a # `condition` that determines how and when the `bindings` are applied. Each # of the `bindings` must contain at least one member. # Corresponds to the JSON property `bindings` # @return [Array] attr_accessor :bindings # `etag` is used for optimistic concurrency control as a way to help # prevent simultaneous updates of a policy from overwriting each other. # It is strongly suggested that systems make use of the `etag` in the # read-modify-write cycle to perform policy updates in order to avoid race # conditions: An `etag` is returned in the response to `getIamPolicy`, and # systems are expected to put that etag in the request to `setIamPolicy` to # ensure that their change will be applied to the same version of the policy. # **Important:** If you use IAM Conditions, you must include the `etag` field # whenever you call `setIamPolicy`. If you omit this field, then IAM allows # you to overwrite a version `3` policy with a version `1` policy, and all of # the conditions in the version `3` policy are lost. # Corresponds to the JSON property `etag` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :etag # Specifies the format of the policy. # Valid values are `0`, `1`, and `3`. Requests that specify an invalid value # are rejected. # Any operation that affects conditional role bindings must specify version # `3`. This requirement applies to the following operations: # * Getting a policy that includes a conditional role binding # * Adding a conditional role binding to a policy # * Changing a conditional role binding in a policy # * Removing any role binding, with or without a condition, from a policy # that includes conditions # **Important:** If you use IAM Conditions, you must include the `etag` field # whenever you call `setIamPolicy`. If you omit this field, then IAM allows # you to overwrite a version `3` policy with a version `1` policy, and all of # the conditions in the version `3` policy are lost. # If a policy does not include any conditions, operations on that policy may # specify any valid version or leave the field unset. # To learn which resources support conditions in their IAM policies, see the # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource- # policies). # Corresponds to the JSON property `version` # @return [Fixnum] attr_accessor :version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @bindings = args[:bindings] if args.key?(:bindings) @etag = args[:etag] if args.key?(:etag) @version = args[:version] if args.key?(:version) end end # Query optimizer configuration. class QueryOptions include Google::Apis::Core::Hashable # An option to control the selection of optimizer version. # This parameter allows individual queries to pick different query # optimizer versions. # Specifying "latest" as a value instructs Cloud Spanner to use the # latest supported query optimizer version. If not specified, Cloud Spanner # uses optimizer version set at the database level options. Any other # positive integer (from the list of supported optimizer versions) # overrides the default optimizer version for query execution. # The list of supported optimizer versions can be queried from # SPANNER_SYS.SUPPORTED_OPTIMIZER_VERSIONS. Executing a SQL statement # with an invalid optimizer version will fail with a syntax error # (`INVALID_ARGUMENT`) status. # See # https://cloud.google.com/spanner/docs/query-optimizer/manage-query-optimizer # for more information on managing the query optimizer. # The `optimizer_version` statement hint has precedence over this setting. # Corresponds to the JSON property `optimizerVersion` # @return [String] attr_accessor :optimizer_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @optimizer_version = args[:optimizer_version] if args.key?(:optimizer_version) end end # Contains an ordered list of nodes appearing in the query plan. class QueryPlan include Google::Apis::Core::Hashable # The nodes in the query plan. Plan nodes are returned in pre-order starting # with the plan root. Each PlanNode's `id` corresponds to its index in # `plan_nodes`. # Corresponds to the JSON property `planNodes` # @return [Array] attr_accessor :plan_nodes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @plan_nodes = args[:plan_nodes] if args.key?(:plan_nodes) end end # Message type to initiate a read-only transaction. class ReadOnly include Google::Apis::Core::Hashable # Executes all reads at a timestamp that is `exact_staleness` # old. The timestamp is chosen soon after the read is started. # Guarantees that all writes that have committed more than the # specified number of seconds ago are visible. Because Cloud Spanner # chooses the exact timestamp, this mode works even if the client's # local clock is substantially skewed from Cloud Spanner commit # timestamps. # Useful for reading at nearby replicas without the distributed # timestamp negotiation overhead of `max_staleness`. # Corresponds to the JSON property `exactStaleness` # @return [String] attr_accessor :exact_staleness # Read data at a timestamp >= `NOW - max_staleness` # seconds. Guarantees that all writes that have committed more # than the specified number of seconds ago are visible. Because # Cloud Spanner chooses the exact timestamp, this mode works even if # the client's local clock is substantially skewed from Cloud Spanner # commit timestamps. # Useful for reading the freshest data available at a nearby # replica, while bounding the possible staleness if the local # replica has fallen behind. # Note that this option can only be used in single-use # transactions. # Corresponds to the JSON property `maxStaleness` # @return [String] attr_accessor :max_staleness # Executes all reads at a timestamp >= `min_read_timestamp`. # This is useful for requesting fresher data than some previous # read, or data that is fresh enough to observe the effects of some # previously committed transaction whose timestamp is known. # Note that this option can only be used in single-use transactions. # A timestamp in RFC3339 UTC \"Zulu\" format, accurate to nanoseconds. # Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `minReadTimestamp` # @return [String] attr_accessor :min_read_timestamp # Executes all reads at the given timestamp. Unlike other modes, # reads at a specific timestamp are repeatable; the same read at # the same timestamp always returns the same data. If the # timestamp is in the future, the read will block until the # specified timestamp, modulo the read's deadline. # Useful for large scale consistent reads such as mapreduces, or # for coordinating many reads against a consistent snapshot of the # data. # A timestamp in RFC3339 UTC \"Zulu\" format, accurate to nanoseconds. # Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `readTimestamp` # @return [String] attr_accessor :read_timestamp # If true, the Cloud Spanner-selected read timestamp is included in # the Transaction message that describes the transaction. # Corresponds to the JSON property `returnReadTimestamp` # @return [Boolean] attr_accessor :return_read_timestamp alias_method :return_read_timestamp?, :return_read_timestamp # Read at a timestamp where all previously committed transactions # are visible. # Corresponds to the JSON property `strong` # @return [Boolean] attr_accessor :strong alias_method :strong?, :strong def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @exact_staleness = args[:exact_staleness] if args.key?(:exact_staleness) @max_staleness = args[:max_staleness] if args.key?(:max_staleness) @min_read_timestamp = args[:min_read_timestamp] if args.key?(:min_read_timestamp) @read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp) @return_read_timestamp = args[:return_read_timestamp] if args.key?(:return_read_timestamp) @strong = args[:strong] if args.key?(:strong) end end # The request for Read and # StreamingRead. class ReadRequest include Google::Apis::Core::Hashable # Required. The columns of table to be returned for each row matching # this request. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # If non-empty, the name of an index on table. This index is # used instead of the table primary key when interpreting key_set # and sorting result rows. See key_set for further information. # Corresponds to the JSON property `index` # @return [String] attr_accessor :index # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All # the keys are expected to be in the same table or index. The keys need # not be sorted in any particular way. # If the same key is specified multiple times in the set (for example # if two ranges, two keys, or a key and a range overlap), Cloud Spanner # behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # If greater than zero, only the first `limit` rows are yielded. If `limit` # is zero, the default is no limit. A limit cannot be specified if # `partition_token` is set. # Corresponds to the JSON property `limit` # @return [Fixnum] attr_accessor :limit # If present, results will be restricted to the specified partition # previously created using PartitionRead(). There must be an exact # match for the values of fields common to this message and the # PartitionReadRequest message used to create this partition_token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token # If this request is resuming a previously interrupted read, # `resume_token` should be copied from the last # PartialResultSet yielded before the interruption. Doing this # enables the new read to resume where the last read left off. The # rest of the request parameters must exactly match the request # that yielded this token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # Required. The name of the table in the database to be read. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @index = args[:index] if args.key?(:index) @key_set = args[:key_set] if args.key?(:key_set) @limit = args[:limit] if args.key?(:limit) @partition_token = args[:partition_token] if args.key?(:partition_token) @resume_token = args[:resume_token] if args.key?(:resume_token) @table = args[:table] if args.key?(:table) @transaction = args[:transaction] if args.key?(:transaction) end end # Message type to initiate a read-write transaction. Currently this # transaction type has no options. class ReadWrite include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # class ReplicaInfo include Google::Apis::Core::Hashable # If true, this location is designated as the default leader location where # leader replicas are placed. See the [region types # documentation](https://cloud.google.com/spanner/docs/instances#region_types) # for more details. # Corresponds to the JSON property `defaultLeaderLocation` # @return [Boolean] attr_accessor :default_leader_location alias_method :default_leader_location?, :default_leader_location # The location of the serving resources, e.g. "us-central1". # Corresponds to the JSON property `location` # @return [String] attr_accessor :location # The type of replica. # Corresponds to the JSON property `type` # @return [String] attr_accessor :type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @default_leader_location = args[:default_leader_location] if args.key?(:default_leader_location) @location = args[:location] if args.key?(:location) @type = args[:type] if args.key?(:type) end end # Metadata type for the long-running operation returned by # RestoreDatabase. class RestoreDatabaseMetadata include Google::Apis::Core::Hashable # Information about a backup. # Corresponds to the JSON property `backupInfo` # @return [Google::Apis::SpannerV1::BackupInfo] attr_accessor :backup_info # The time at which cancellation of this operation was received. # Operations.CancelOperation # starts asynchronous cancellation on a long-running operation. The server # makes a best effort to cancel the operation, but success is not guaranteed. # Clients can use # Operations.GetOperation or # other methods to check whether the cancellation succeeded or whether the # operation completed despite cancellation. On successful cancellation, # the operation is not deleted; instead, it becomes an operation with # an Operation.error value with a # google.rpc.Status.code of 1, corresponding to `Code.CANCELLED`. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # Name of the database being created and restored to. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # If exists, the name of the long-running operation that will be used to # track the post-restore optimization process to optimize the performance of # the restored database, and remove the dependency on the restore source. # The name is of the form # `projects//instances//databases//operations/< # operation>` # where the is the name of database being created and restored to. # The metadata type of the long-running operation is # OptimizeRestoredDatabaseMetadata. This long-running operation will be # automatically created by the system after the RestoreDatabase long-running # operation completes successfully. This operation will not be created if the # restore was not successful. # Corresponds to the JSON property `optimizeDatabaseOperationName` # @return [String] attr_accessor :optimize_database_operation_name # Encapsulates progress related information for a Cloud Spanner long # running operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress # The type of the restore source. # Corresponds to the JSON property `sourceType` # @return [String] attr_accessor :source_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup_info = args[:backup_info] if args.key?(:backup_info) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @name = args[:name] if args.key?(:name) @optimize_database_operation_name = args[:optimize_database_operation_name] if args.key?(:optimize_database_operation_name) @progress = args[:progress] if args.key?(:progress) @source_type = args[:source_type] if args.key?(:source_type) end end # The request for # RestoreDatabase. class RestoreDatabaseRequest include Google::Apis::Core::Hashable # Name of the backup from which to restore. Values are of the form # `projects//instances//backups/`. # Corresponds to the JSON property `backup` # @return [String] attr_accessor :backup # Required. The id of the database to create and restore to. This # database must not already exist. The `database_id` appended to # `parent` forms the full database name of the form # `projects//instances//databases/`. # Corresponds to the JSON property `databaseId` # @return [String] attr_accessor :database_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup = args[:backup] if args.key?(:backup) @database_id = args[:database_id] if args.key?(:database_id) end end # Information about the database restore. class RestoreInfo include Google::Apis::Core::Hashable # Information about a backup. # Corresponds to the JSON property `backupInfo` # @return [Google::Apis::SpannerV1::BackupInfo] attr_accessor :backup_info # The type of the restore source. # Corresponds to the JSON property `sourceType` # @return [String] attr_accessor :source_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup_info = args[:backup_info] if args.key?(:backup_info) @source_type = args[:source_type] if args.key?(:source_type) end end # Results from Read or # ExecuteSql. class ResultSet include Google::Apis::Core::Hashable # Metadata about a ResultSet or PartialResultSet. # Corresponds to the JSON property `metadata` # @return [Google::Apis::SpannerV1::ResultSetMetadata] attr_accessor :metadata # Each element in `rows` is a row whose format is defined by # metadata.row_type. The ith element # in each row matches the ith field in # metadata.row_type. Elements are # encoded based on type as described # here. # Corresponds to the JSON property `rows` # @return [Array>] attr_accessor :rows # Additional statistics about a ResultSet or PartialResultSet. # Corresponds to the JSON property `stats` # @return [Google::Apis::SpannerV1::ResultSetStats] attr_accessor :stats def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @metadata = args[:metadata] if args.key?(:metadata) @rows = args[:rows] if args.key?(:rows) @stats = args[:stats] if args.key?(:stats) end end # Metadata about a ResultSet or PartialResultSet. class ResultSetMetadata include Google::Apis::Core::Hashable # `StructType` defines the fields of a STRUCT type. # Corresponds to the JSON property `rowType` # @return [Google::Apis::SpannerV1::StructType] attr_accessor :row_type # A transaction. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::Transaction] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @row_type = args[:row_type] if args.key?(:row_type) @transaction = args[:transaction] if args.key?(:transaction) end end # Additional statistics about a ResultSet or PartialResultSet. class ResultSetStats include Google::Apis::Core::Hashable # Contains an ordered list of nodes appearing in the query plan. # Corresponds to the JSON property `queryPlan` # @return [Google::Apis::SpannerV1::QueryPlan] attr_accessor :query_plan # Aggregated statistics from the execution of the query. Only present when # the query is profiled. For example, a query could return the statistics as # follows: # ` # "rows_returned": "3", # "elapsed_time": "1.22 secs", # "cpu_time": "1.19 secs" # ` # Corresponds to the JSON property `queryStats` # @return [Hash] attr_accessor :query_stats # Standard DML returns an exact count of rows that were modified. # Corresponds to the JSON property `rowCountExact` # @return [Fixnum] attr_accessor :row_count_exact # Partitioned DML does not offer exactly-once semantics, so it # returns a lower bound of the rows modified. # Corresponds to the JSON property `rowCountLowerBound` # @return [Fixnum] attr_accessor :row_count_lower_bound def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @query_plan = args[:query_plan] if args.key?(:query_plan) @query_stats = args[:query_stats] if args.key?(:query_stats) @row_count_exact = args[:row_count_exact] if args.key?(:row_count_exact) @row_count_lower_bound = args[:row_count_lower_bound] if args.key?(:row_count_lower_bound) end end # The request for Rollback. class RollbackRequest include Google::Apis::Core::Hashable # Required. The transaction to roll back. # Corresponds to the JSON property `transactionId` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :transaction_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @transaction_id = args[:transaction_id] if args.key?(:transaction_id) end end # A session in the Cloud Spanner API. class Session include Google::Apis::Core::Hashable # Output only. The approximate timestamp when the session is last used. It is # typically earlier than the actual last use time. # Corresponds to the JSON property `approximateLastUseTime` # @return [String] attr_accessor :approximate_last_use_time # Output only. The timestamp when the session is created. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # The labels for the session. # * Label keys must be between 1 and 63 characters long and must conform to # the following regular expression: `[a-z]([-a-z0-9]*[a-z0-9])?`. # * Label values must be between 0 and 63 characters long and must conform # to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. # * No more than 64 labels can be associated with a given session. # See https://goo.gl/xmQnxf for more information on and examples of labels. # Corresponds to the JSON property `labels` # @return [Hash] attr_accessor :labels # Output only. The name of the session. This is always system-assigned. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @approximate_last_use_time = args[:approximate_last_use_time] if args.key?(:approximate_last_use_time) @create_time = args[:create_time] if args.key?(:create_time) @labels = args[:labels] if args.key?(:labels) @name = args[:name] if args.key?(:name) end end # Request message for `SetIamPolicy` method. class SetIamPolicyRequest include Google::Apis::Core::Hashable # An Identity and Access Management (IAM) policy, which specifies access # controls for Google Cloud resources. # A `Policy` is a collection of `bindings`. A `binding` binds one or more # `members` to a single `role`. Members can be user accounts, service accounts, # Google groups, and domains (such as G Suite). A `role` is a named list of # permissions; each `role` can be an IAM predefined role or a user-created # custom role. # For some types of Google Cloud resources, a `binding` can also specify a # `condition`, which is a logical expression that allows access to a resource # only if the expression evaluates to `true`. A condition can add constraints # based on attributes of the request, the resource, or both. To learn which # resources support conditions in their IAM policies, see the # [IAM documentation](https://cloud.google.com/iam/help/conditions/resource- # policies). # **JSON example:** # ` # "bindings": [ # ` # "role": "roles/resourcemanager.organizationAdmin", # "members": [ # "user:mike@example.com", # "group:admins@example.com", # "domain:google.com", # "serviceAccount:my-project-id@appspot.gserviceaccount.com" # ] # `, # ` # "role": "roles/resourcemanager.organizationViewer", # "members": [ # "user:eve@example.com" # ], # "condition": ` # "title": "expirable access", # "description": "Does not grant access after Sep 2020", # "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z') # ", # ` # ` # ], # "etag": "BwWWja0YfJA=", # "version": 3 # ` # **YAML example:** # bindings: # - members: # - user:mike@example.com # - group:admins@example.com # - domain:google.com # - serviceAccount:my-project-id@appspot.gserviceaccount.com # role: roles/resourcemanager.organizationAdmin # - members: # - user:eve@example.com # role: roles/resourcemanager.organizationViewer # condition: # title: expirable access # description: Does not grant access after Sep 2020 # expression: request.time < timestamp('2020-10-01T00:00:00.000Z') # - etag: BwWWja0YfJA= # - version: 3 # For a description of IAM and its features, see the # [IAM documentation](https://cloud.google.com/iam/docs/). # Corresponds to the JSON property `policy` # @return [Google::Apis::SpannerV1::Policy] attr_accessor :policy def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @policy = args[:policy] if args.key?(:policy) end end # Condensed representation of a node and its subtree. Only present for # `SCALAR` PlanNode(s). class ShortRepresentation include Google::Apis::Core::Hashable # A string representation of the expression subtree rooted at this node. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description # A mapping of (subquery variable name) -> (subquery node id) for cases # where the `description` string of this node references a `SCALAR` # subquery contained in the expression subtree rooted at this node. The # referenced `SCALAR` subquery may not necessarily be a direct child of # this node. # Corresponds to the JSON property `subqueries` # @return [Hash] attr_accessor :subqueries def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @description = args[:description] if args.key?(:description) @subqueries = args[:subqueries] if args.key?(:subqueries) end end # A single DML statement. class Statement include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type # from a JSON value. For example, values of type `BYTES` and values # of type `STRING` both appear in params as JSON strings. # In these cases, `param_types` can be used to specify the exact # SQL type for some or all of the SQL statement parameters. See the # definition of Type for more information # about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the DML string. # A parameter placeholder consists of the `@` character followed by the # parameter name (for example, `@firstName`). Parameter names can contain # letters, numbers, and underscores. # Parameters can appear anywhere that a literal value is expected. The # same parameter name can be used more than once, for example: # `"WHERE id > @msg_id AND id < @msg_id + 100"` # It is an error to execute a SQL statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # Required. The DML string. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @sql = args[:sql] if args.key?(:sql) end end # The `Status` type defines a logical error model that is suitable for # different programming environments, including REST APIs and RPC APIs. It is # used by [gRPC](https://github.com/grpc). Each `Status` message contains # three pieces of data: error code, error message, and error details. # You can find out more about this error model and how to work with it in the # [API Design Guide](https://cloud.google.com/apis/design/errors). class Status include Google::Apis::Core::Hashable # The status code, which should be an enum value of google.rpc.Code. # Corresponds to the JSON property `code` # @return [Fixnum] attr_accessor :code # A list of messages that carry the error details. There is a common set of # message types for APIs to use. # Corresponds to the JSON property `details` # @return [Array>] attr_accessor :details # A developer-facing error message, which should be in English. Any # user-facing error message should be localized and sent in the # google.rpc.Status.details field, or localized by the client. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @code = args[:code] if args.key?(:code) @details = args[:details] if args.key?(:details) @message = args[:message] if args.key?(:message) end end # `StructType` defines the fields of a STRUCT type. class StructType include Google::Apis::Core::Hashable # The list of fields that make up this struct. Order is # significant, because values of this struct type are represented as # lists, where the order of field values matches the order of # fields in the StructType. In turn, the order of fields # matches the order of columns in a read request, or the order of # fields in the `SELECT` clause of a query. # Corresponds to the JSON property `fields` # @return [Array] attr_accessor :fields def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @fields = args[:fields] if args.key?(:fields) end end # Request message for `TestIamPermissions` method. class TestIamPermissionsRequest include Google::Apis::Core::Hashable # REQUIRED: The set of permissions to check for 'resource'. # Permissions with wildcards (such as '*', 'spanner.*', 'spanner.instances.*') # are not allowed. # Corresponds to the JSON property `permissions` # @return [Array] attr_accessor :permissions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @permissions = args[:permissions] if args.key?(:permissions) end end # Response message for `TestIamPermissions` method. class TestIamPermissionsResponse include Google::Apis::Core::Hashable # A subset of `TestPermissionsRequest.permissions` that the caller is # allowed. # Corresponds to the JSON property `permissions` # @return [Array] attr_accessor :permissions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @permissions = args[:permissions] if args.key?(:permissions) end end # A transaction. class Transaction include Google::Apis::Core::Hashable # `id` may be used to identify the transaction in subsequent # Read, # ExecuteSql, # Commit, or # Rollback calls. # Single-use read-only transactions do not have IDs, because # single-use transactions do not support multiple requests. # Corresponds to the JSON property `id` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :id # For snapshot read-only transactions, the read timestamp chosen # for the transaction. Not returned by default: see # TransactionOptions.ReadOnly.return_read_timestamp. # A timestamp in RFC3339 UTC \"Zulu\" format, accurate to nanoseconds. # Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `readTimestamp` # @return [String] attr_accessor :read_timestamp def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @id = args[:id] if args.key?(:id) @read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp) end end # # Transactions # Each session can have at most one active transaction at a time (note that # standalone reads and queries use a transaction internally and do count # towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. # It is not necessary to create a new session for each transaction. # # Transaction Modes # Cloud Spanner supports three transaction modes: # 1. Locking read-write. This type of transaction is the only way # to write data into Cloud Spanner. These transactions rely on # pessimistic locking and, if necessary, two-phase commit. # Locking read-write transactions may abort, requiring the # application to retry. # 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow # writes. Snapshot read-only transactions can be configured to # read at timestamps in the past. Snapshot read-only # transactions do not need to be committed. # 3. Partitioned DML. This type of transaction is used to execute # a single Partitioned DML statement. Partitioned DML partitions # the key space and runs the DML statement over each partition # in parallel using separate, internal transactions that commit # independently. Partitioned DML transactions do not need to be # committed. # For transactions that only read, snapshot read-only transactions # provide simpler semantics and are almost always faster. In # particular, read-only transactions do not take locks, so they do # not conflict with read-write transactions. As a consequence of not # taking locks, they also do not abort, so retry loops are not needed. # Transactions may only read/write data in a single database. They # may, however, read/write data in different tables within that # database. # ## Locking Read-Write Transactions # Locking transactions may be used to atomically read-modify-write # data anywhere in a database. This type of transaction is externally # consistent. # Clients should attempt to minimize the amount of time a transaction # is active. Faster transactions commit with higher probability # and cause less contention. Cloud Spanner attempts to keep read locks # active as long as the transaction continues to do reads, and the # transaction has not been terminated by # Commit or # Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a # transaction's locks and abort it. # Conceptually, a read-write transaction consists of zero or more # reads or SQL statements followed by # Commit. At any time before # Commit, the client can send a # Rollback request to abort the # transaction. # ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired # are still valid at commit time, and it is able to acquire write # locks for all writes. Cloud Spanner can abort the transaction for any # reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees # that the transaction has not modified any user data in Cloud Spanner. # Unless the transaction commits, Cloud Spanner makes no guarantees about # how long the transaction's locks were held for. It is an error to # use Cloud Spanner locks for any sort of mutual exclusion other than # between Cloud Spanner transactions themselves. # ### Retrying Aborted Transactions # When a transaction aborts, the application can choose to retry the # whole transaction again. To maximize the chances of successfully # committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock # priority increases with each consecutive abort, meaning that each # attempt has a slightly better chance of success than the previous. # Under some circumstances (e.g., many transactions attempting to # modify the same row(s)), a transaction can abort many times in a # short period before successfully committing. Thus, it is not a good # idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent # retrying. # ### Idle Transactions # A transaction is considered idle if it has no outstanding reads or # SQL queries and has not started a read or SQL query within the last 10 # seconds. Idle transactions can be aborted by Cloud Spanner so that they # don't hold on to locks indefinitely. In that case, the commit will # fail with error `ABORTED`. # If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the # transaction from becoming idle. # ## Snapshot Read-Only Transactions # Snapshot read-only transactions provides a simpler method than # locking read-write transactions for doing several consistent # reads. However, this type of transaction does not support writes. # Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that # timestamp. Since they do not acquire locks, they do not block # concurrent read-write transactions. # Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read # timestamp is garbage collected; however, the default garbage # collection policy is generous enough that most applications do not # need to worry about this in practice. # Snapshot read-only transactions do not need to call # Commit or # Rollback (and in fact are not # permitted to do so). # To execute a snapshot transaction, the client specifies a timestamp # bound, which tells Cloud Spanner how to choose a read timestamp. # The types of timestamp bound are: # - Strong (the default). # - Bounded staleness. # - Exact staleness. # If the Cloud Spanner database to be read is geographically distributed, # stale read-only transactions can execute more quickly than strong # or read-write transaction, because they are able to execute far # from the leader replica. # Each type of timestamp bound is discussed in detail below. # ### Strong # Strong reads are guaranteed to see the effects of all transactions # that have committed before the start of the read. Furthermore, all # rows yielded by a single read are consistent with each other -- if # any part of the read observes a transaction, all parts of the read # see the transaction. # Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are # concurrent writes. If consistency across reads is required, the # reads should be executed within a transaction or at an exact read # timestamp. # See TransactionOptions.ReadOnly.strong. # ### Exact Staleness # These timestamp bounds execute reads at a user-specified # timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe # modifications done by all transactions with a commit timestamp <= # the read timestamp, and observe none of the modifications done by # transactions with a larger commit timestamp. They will block until # all conflicting transactions that may be assigned commit timestamps # <= the read timestamp have finished. # The timestamp can either be expressed as an absolute Cloud Spanner commit # timestamp or a staleness relative to the current time. # These modes do not require a "negotiation phase" to pick a # timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, # boundedly stale reads usually return fresher results. # See TransactionOptions.ReadOnly.read_timestamp and # TransactionOptions.ReadOnly.exact_staleness. # ### Bounded Staleness # Bounded staleness modes allow Cloud Spanner to pick the read timestamp, # subject to a user-provided staleness bound. Cloud Spanner chooses the # newest timestamp within the staleness bound that allows execution # of the reads at the closest available replica without blocking. # All rows yielded are consistent with each other -- if any part of # the read observes a transaction, all parts of the read see the # transaction. Boundedly stale reads are not repeatable: two stale # reads, even if they use the same staleness bound, can execute at # different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase # negotiates a timestamp among all replicas needed to serve the # read. In the second phase, reads are executed at the negotiated # timestamp. # As a result of the two phase execution, bounded staleness reads are # usually a little slower than comparable exact staleness # reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. # Because the timestamp negotiation requires up-front knowledge of # which rows will be read, it can only be used with single-use # read-only transactions. # See TransactionOptions.ReadOnly.max_staleness and # TransactionOptions.ReadOnly.min_read_timestamp. # ### Old Read Timestamps and Garbage Collection # Cloud Spanner continuously garbage collects deleted and overwritten data # in the background to reclaim storage space. This process is known # as "version GC". By default, version GC reclaims versions after they # are one hour old. Because of this, Cloud Spanner cannot perform reads # at read timestamps more than one hour in the past. This # restriction also applies to in-progress reads and/or SQL queries whose # timestamp become too old while executing. Reads and SQL queries with # too-old read timestamps fail with the error `FAILED_PRECONDITION`. # ## Partitioned DML Transactions # Partitioned DML transactions are used to execute DML statements with a # different execution strategy that provides different, and often better, # scalability properties for large, table-wide operations than DML in a # ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, # should prefer using ReadWrite transactions. # Partitioned DML partitions the keyspace and runs the DML statement on each # partition in separate, internal transactions. These transactions commit # automatically when complete, and run independently from one another. # To reduce lock contention, this execution strategy only acquires read locks # on rows that match the WHERE clause of the statement. Additionally, the # smaller per-partition transactions hold locks for less time. # That said, Partitioned DML is not a drop-in replacement for standard DML used # in ReadWrite transactions. # - The DML statement must be fully-partitionable. Specifically, the statement # must be expressible as the union of many statements which each access only # a single row of the table. # - The statement is not applied atomically to all rows of the table. Rather, # the statement is applied atomically to partitions of the table, in # independent transactions. Secondary index rows are updated atomically # with the base table rows. # - Partitioned DML does not guarantee exactly-once execution semantics # against a partition. The statement will be applied at least once to each # partition. It is strongly recommended that the DML statement should be # idempotent to avoid unexpected results. For instance, it is potentially # dangerous to run a statement such as # `UPDATE table SET column = column + 1` as it could be run multiple times # against some rows. # - The partitions are committed automatically - there is no support for # Commit or Rollback. If the call returns an error, or if the client issuing # the ExecuteSql call dies, it is possible that some rows had the statement # executed on them successfully. It is also possible that statement was # never executed against other rows. # - Partitioned DML transactions may only contain the execution of a single # DML statement via ExecuteSql or ExecuteStreamingSql. # - If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the # operation is stopped at that point and an error is returned. It is # possible that at this point, some partitions have been committed (or even # committed multiple times), and other partitions have not been run at all. # Given the above, Partitioned DML is good fit for large, database-wide, # operations that are idempotent, such as deleting old rows from a very large # table. class TransactionOptions include Google::Apis::Core::Hashable # Message type to initiate a Partitioned DML transaction. # Corresponds to the JSON property `partitionedDml` # @return [Google::Apis::SpannerV1::PartitionedDml] attr_accessor :partitioned_dml # Message type to initiate a read-only transaction. # Corresponds to the JSON property `readOnly` # @return [Google::Apis::SpannerV1::ReadOnly] attr_accessor :read_only # Message type to initiate a read-write transaction. Currently this # transaction type has no options. # Corresponds to the JSON property `readWrite` # @return [Google::Apis::SpannerV1::ReadWrite] attr_accessor :read_write def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partitioned_dml = args[:partitioned_dml] if args.key?(:partitioned_dml) @read_only = args[:read_only] if args.key?(:read_only) @read_write = args[:read_write] if args.key?(:read_write) end end # This message is used to select the transaction in which a # Read or # ExecuteSql call runs. # See TransactionOptions for more information about transactions. class TransactionSelector include Google::Apis::Core::Hashable # # Transactions # Each session can have at most one active transaction at a time (note that # standalone reads and queries use a transaction internally and do count # towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. # It is not necessary to create a new session for each transaction. # # Transaction Modes # Cloud Spanner supports three transaction modes: # 1. Locking read-write. This type of transaction is the only way # to write data into Cloud Spanner. These transactions rely on # pessimistic locking and, if necessary, two-phase commit. # Locking read-write transactions may abort, requiring the # application to retry. # 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow # writes. Snapshot read-only transactions can be configured to # read at timestamps in the past. Snapshot read-only # transactions do not need to be committed. # 3. Partitioned DML. This type of transaction is used to execute # a single Partitioned DML statement. Partitioned DML partitions # the key space and runs the DML statement over each partition # in parallel using separate, internal transactions that commit # independently. Partitioned DML transactions do not need to be # committed. # For transactions that only read, snapshot read-only transactions # provide simpler semantics and are almost always faster. In # particular, read-only transactions do not take locks, so they do # not conflict with read-write transactions. As a consequence of not # taking locks, they also do not abort, so retry loops are not needed. # Transactions may only read/write data in a single database. They # may, however, read/write data in different tables within that # database. # ## Locking Read-Write Transactions # Locking transactions may be used to atomically read-modify-write # data anywhere in a database. This type of transaction is externally # consistent. # Clients should attempt to minimize the amount of time a transaction # is active. Faster transactions commit with higher probability # and cause less contention. Cloud Spanner attempts to keep read locks # active as long as the transaction continues to do reads, and the # transaction has not been terminated by # Commit or # Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a # transaction's locks and abort it. # Conceptually, a read-write transaction consists of zero or more # reads or SQL statements followed by # Commit. At any time before # Commit, the client can send a # Rollback request to abort the # transaction. # ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired # are still valid at commit time, and it is able to acquire write # locks for all writes. Cloud Spanner can abort the transaction for any # reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees # that the transaction has not modified any user data in Cloud Spanner. # Unless the transaction commits, Cloud Spanner makes no guarantees about # how long the transaction's locks were held for. It is an error to # use Cloud Spanner locks for any sort of mutual exclusion other than # between Cloud Spanner transactions themselves. # ### Retrying Aborted Transactions # When a transaction aborts, the application can choose to retry the # whole transaction again. To maximize the chances of successfully # committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock # priority increases with each consecutive abort, meaning that each # attempt has a slightly better chance of success than the previous. # Under some circumstances (e.g., many transactions attempting to # modify the same row(s)), a transaction can abort many times in a # short period before successfully committing. Thus, it is not a good # idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent # retrying. # ### Idle Transactions # A transaction is considered idle if it has no outstanding reads or # SQL queries and has not started a read or SQL query within the last 10 # seconds. Idle transactions can be aborted by Cloud Spanner so that they # don't hold on to locks indefinitely. In that case, the commit will # fail with error `ABORTED`. # If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the # transaction from becoming idle. # ## Snapshot Read-Only Transactions # Snapshot read-only transactions provides a simpler method than # locking read-write transactions for doing several consistent # reads. However, this type of transaction does not support writes. # Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that # timestamp. Since they do not acquire locks, they do not block # concurrent read-write transactions. # Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read # timestamp is garbage collected; however, the default garbage # collection policy is generous enough that most applications do not # need to worry about this in practice. # Snapshot read-only transactions do not need to call # Commit or # Rollback (and in fact are not # permitted to do so). # To execute a snapshot transaction, the client specifies a timestamp # bound, which tells Cloud Spanner how to choose a read timestamp. # The types of timestamp bound are: # - Strong (the default). # - Bounded staleness. # - Exact staleness. # If the Cloud Spanner database to be read is geographically distributed, # stale read-only transactions can execute more quickly than strong # or read-write transaction, because they are able to execute far # from the leader replica. # Each type of timestamp bound is discussed in detail below. # ### Strong # Strong reads are guaranteed to see the effects of all transactions # that have committed before the start of the read. Furthermore, all # rows yielded by a single read are consistent with each other -- if # any part of the read observes a transaction, all parts of the read # see the transaction. # Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are # concurrent writes. If consistency across reads is required, the # reads should be executed within a transaction or at an exact read # timestamp. # See TransactionOptions.ReadOnly.strong. # ### Exact Staleness # These timestamp bounds execute reads at a user-specified # timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe # modifications done by all transactions with a commit timestamp <= # the read timestamp, and observe none of the modifications done by # transactions with a larger commit timestamp. They will block until # all conflicting transactions that may be assigned commit timestamps # <= the read timestamp have finished. # The timestamp can either be expressed as an absolute Cloud Spanner commit # timestamp or a staleness relative to the current time. # These modes do not require a "negotiation phase" to pick a # timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, # boundedly stale reads usually return fresher results. # See TransactionOptions.ReadOnly.read_timestamp and # TransactionOptions.ReadOnly.exact_staleness. # ### Bounded Staleness # Bounded staleness modes allow Cloud Spanner to pick the read timestamp, # subject to a user-provided staleness bound. Cloud Spanner chooses the # newest timestamp within the staleness bound that allows execution # of the reads at the closest available replica without blocking. # All rows yielded are consistent with each other -- if any part of # the read observes a transaction, all parts of the read see the # transaction. Boundedly stale reads are not repeatable: two stale # reads, even if they use the same staleness bound, can execute at # different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase # negotiates a timestamp among all replicas needed to serve the # read. In the second phase, reads are executed at the negotiated # timestamp. # As a result of the two phase execution, bounded staleness reads are # usually a little slower than comparable exact staleness # reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. # Because the timestamp negotiation requires up-front knowledge of # which rows will be read, it can only be used with single-use # read-only transactions. # See TransactionOptions.ReadOnly.max_staleness and # TransactionOptions.ReadOnly.min_read_timestamp. # ### Old Read Timestamps and Garbage Collection # Cloud Spanner continuously garbage collects deleted and overwritten data # in the background to reclaim storage space. This process is known # as "version GC". By default, version GC reclaims versions after they # are one hour old. Because of this, Cloud Spanner cannot perform reads # at read timestamps more than one hour in the past. This # restriction also applies to in-progress reads and/or SQL queries whose # timestamp become too old while executing. Reads and SQL queries with # too-old read timestamps fail with the error `FAILED_PRECONDITION`. # ## Partitioned DML Transactions # Partitioned DML transactions are used to execute DML statements with a # different execution strategy that provides different, and often better, # scalability properties for large, table-wide operations than DML in a # ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, # should prefer using ReadWrite transactions. # Partitioned DML partitions the keyspace and runs the DML statement on each # partition in separate, internal transactions. These transactions commit # automatically when complete, and run independently from one another. # To reduce lock contention, this execution strategy only acquires read locks # on rows that match the WHERE clause of the statement. Additionally, the # smaller per-partition transactions hold locks for less time. # That said, Partitioned DML is not a drop-in replacement for standard DML used # in ReadWrite transactions. # - The DML statement must be fully-partitionable. Specifically, the statement # must be expressible as the union of many statements which each access only # a single row of the table. # - The statement is not applied atomically to all rows of the table. Rather, # the statement is applied atomically to partitions of the table, in # independent transactions. Secondary index rows are updated atomically # with the base table rows. # - Partitioned DML does not guarantee exactly-once execution semantics # against a partition. The statement will be applied at least once to each # partition. It is strongly recommended that the DML statement should be # idempotent to avoid unexpected results. For instance, it is potentially # dangerous to run a statement such as # `UPDATE table SET column = column + 1` as it could be run multiple times # against some rows. # - The partitions are committed automatically - there is no support for # Commit or Rollback. If the call returns an error, or if the client issuing # the ExecuteSql call dies, it is possible that some rows had the statement # executed on them successfully. It is also possible that statement was # never executed against other rows. # - Partitioned DML transactions may only contain the execution of a single # DML statement via ExecuteSql or ExecuteStreamingSql. # - If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the # operation is stopped at that point and an error is returned. It is # possible that at this point, some partitions have been committed (or even # committed multiple times), and other partitions have not been run at all. # Given the above, Partitioned DML is good fit for large, database-wide, # operations that are idempotent, such as deleting old rows from a very large # table. # Corresponds to the JSON property `begin` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :begin # Execute the read or SQL query in a previously-started transaction. # Corresponds to the JSON property `id` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :id # # Transactions # Each session can have at most one active transaction at a time (note that # standalone reads and queries use a transaction internally and do count # towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. # It is not necessary to create a new session for each transaction. # # Transaction Modes # Cloud Spanner supports three transaction modes: # 1. Locking read-write. This type of transaction is the only way # to write data into Cloud Spanner. These transactions rely on # pessimistic locking and, if necessary, two-phase commit. # Locking read-write transactions may abort, requiring the # application to retry. # 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow # writes. Snapshot read-only transactions can be configured to # read at timestamps in the past. Snapshot read-only # transactions do not need to be committed. # 3. Partitioned DML. This type of transaction is used to execute # a single Partitioned DML statement. Partitioned DML partitions # the key space and runs the DML statement over each partition # in parallel using separate, internal transactions that commit # independently. Partitioned DML transactions do not need to be # committed. # For transactions that only read, snapshot read-only transactions # provide simpler semantics and are almost always faster. In # particular, read-only transactions do not take locks, so they do # not conflict with read-write transactions. As a consequence of not # taking locks, they also do not abort, so retry loops are not needed. # Transactions may only read/write data in a single database. They # may, however, read/write data in different tables within that # database. # ## Locking Read-Write Transactions # Locking transactions may be used to atomically read-modify-write # data anywhere in a database. This type of transaction is externally # consistent. # Clients should attempt to minimize the amount of time a transaction # is active. Faster transactions commit with higher probability # and cause less contention. Cloud Spanner attempts to keep read locks # active as long as the transaction continues to do reads, and the # transaction has not been terminated by # Commit or # Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a # transaction's locks and abort it. # Conceptually, a read-write transaction consists of zero or more # reads or SQL statements followed by # Commit. At any time before # Commit, the client can send a # Rollback request to abort the # transaction. # ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired # are still valid at commit time, and it is able to acquire write # locks for all writes. Cloud Spanner can abort the transaction for any # reason. If a commit attempt returns `ABORTED`, Cloud Spanner guarantees # that the transaction has not modified any user data in Cloud Spanner. # Unless the transaction commits, Cloud Spanner makes no guarantees about # how long the transaction's locks were held for. It is an error to # use Cloud Spanner locks for any sort of mutual exclusion other than # between Cloud Spanner transactions themselves. # ### Retrying Aborted Transactions # When a transaction aborts, the application can choose to retry the # whole transaction again. To maximize the chances of successfully # committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock # priority increases with each consecutive abort, meaning that each # attempt has a slightly better chance of success than the previous. # Under some circumstances (e.g., many transactions attempting to # modify the same row(s)), a transaction can abort many times in a # short period before successfully committing. Thus, it is not a good # idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent # retrying. # ### Idle Transactions # A transaction is considered idle if it has no outstanding reads or # SQL queries and has not started a read or SQL query within the last 10 # seconds. Idle transactions can be aborted by Cloud Spanner so that they # don't hold on to locks indefinitely. In that case, the commit will # fail with error `ABORTED`. # If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the # transaction from becoming idle. # ## Snapshot Read-Only Transactions # Snapshot read-only transactions provides a simpler method than # locking read-write transactions for doing several consistent # reads. However, this type of transaction does not support writes. # Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that # timestamp. Since they do not acquire locks, they do not block # concurrent read-write transactions. # Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read # timestamp is garbage collected; however, the default garbage # collection policy is generous enough that most applications do not # need to worry about this in practice. # Snapshot read-only transactions do not need to call # Commit or # Rollback (and in fact are not # permitted to do so). # To execute a snapshot transaction, the client specifies a timestamp # bound, which tells Cloud Spanner how to choose a read timestamp. # The types of timestamp bound are: # - Strong (the default). # - Bounded staleness. # - Exact staleness. # If the Cloud Spanner database to be read is geographically distributed, # stale read-only transactions can execute more quickly than strong # or read-write transaction, because they are able to execute far # from the leader replica. # Each type of timestamp bound is discussed in detail below. # ### Strong # Strong reads are guaranteed to see the effects of all transactions # that have committed before the start of the read. Furthermore, all # rows yielded by a single read are consistent with each other -- if # any part of the read observes a transaction, all parts of the read # see the transaction. # Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are # concurrent writes. If consistency across reads is required, the # reads should be executed within a transaction or at an exact read # timestamp. # See TransactionOptions.ReadOnly.strong. # ### Exact Staleness # These timestamp bounds execute reads at a user-specified # timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe # modifications done by all transactions with a commit timestamp <= # the read timestamp, and observe none of the modifications done by # transactions with a larger commit timestamp. They will block until # all conflicting transactions that may be assigned commit timestamps # <= the read timestamp have finished. # The timestamp can either be expressed as an absolute Cloud Spanner commit # timestamp or a staleness relative to the current time. # These modes do not require a "negotiation phase" to pick a # timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, # boundedly stale reads usually return fresher results. # See TransactionOptions.ReadOnly.read_timestamp and # TransactionOptions.ReadOnly.exact_staleness. # ### Bounded Staleness # Bounded staleness modes allow Cloud Spanner to pick the read timestamp, # subject to a user-provided staleness bound. Cloud Spanner chooses the # newest timestamp within the staleness bound that allows execution # of the reads at the closest available replica without blocking. # All rows yielded are consistent with each other -- if any part of # the read observes a transaction, all parts of the read see the # transaction. Boundedly stale reads are not repeatable: two stale # reads, even if they use the same staleness bound, can execute at # different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase # negotiates a timestamp among all replicas needed to serve the # read. In the second phase, reads are executed at the negotiated # timestamp. # As a result of the two phase execution, bounded staleness reads are # usually a little slower than comparable exact staleness # reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. # Because the timestamp negotiation requires up-front knowledge of # which rows will be read, it can only be used with single-use # read-only transactions. # See TransactionOptions.ReadOnly.max_staleness and # TransactionOptions.ReadOnly.min_read_timestamp. # ### Old Read Timestamps and Garbage Collection # Cloud Spanner continuously garbage collects deleted and overwritten data # in the background to reclaim storage space. This process is known # as "version GC". By default, version GC reclaims versions after they # are one hour old. Because of this, Cloud Spanner cannot perform reads # at read timestamps more than one hour in the past. This # restriction also applies to in-progress reads and/or SQL queries whose # timestamp become too old while executing. Reads and SQL queries with # too-old read timestamps fail with the error `FAILED_PRECONDITION`. # ## Partitioned DML Transactions # Partitioned DML transactions are used to execute DML statements with a # different execution strategy that provides different, and often better, # scalability properties for large, table-wide operations than DML in a # ReadWrite transaction. Smaller scoped statements, such as an OLTP workload, # should prefer using ReadWrite transactions. # Partitioned DML partitions the keyspace and runs the DML statement on each # partition in separate, internal transactions. These transactions commit # automatically when complete, and run independently from one another. # To reduce lock contention, this execution strategy only acquires read locks # on rows that match the WHERE clause of the statement. Additionally, the # smaller per-partition transactions hold locks for less time. # That said, Partitioned DML is not a drop-in replacement for standard DML used # in ReadWrite transactions. # - The DML statement must be fully-partitionable. Specifically, the statement # must be expressible as the union of many statements which each access only # a single row of the table. # - The statement is not applied atomically to all rows of the table. Rather, # the statement is applied atomically to partitions of the table, in # independent transactions. Secondary index rows are updated atomically # with the base table rows. # - Partitioned DML does not guarantee exactly-once execution semantics # against a partition. The statement will be applied at least once to each # partition. It is strongly recommended that the DML statement should be # idempotent to avoid unexpected results. For instance, it is potentially # dangerous to run a statement such as # `UPDATE table SET column = column + 1` as it could be run multiple times # against some rows. # - The partitions are committed automatically - there is no support for # Commit or Rollback. If the call returns an error, or if the client issuing # the ExecuteSql call dies, it is possible that some rows had the statement # executed on them successfully. It is also possible that statement was # never executed against other rows. # - Partitioned DML transactions may only contain the execution of a single # DML statement via ExecuteSql or ExecuteStreamingSql. # - If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the # operation is stopped at that point and an error is returned. It is # possible that at this point, some partitions have been committed (or even # committed multiple times), and other partitions have not been run at all. # Given the above, Partitioned DML is good fit for large, database-wide, # operations that are idempotent, such as deleting old rows from a very large # table. # Corresponds to the JSON property `singleUse` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :single_use def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @begin = args[:begin] if args.key?(:begin) @id = args[:id] if args.key?(:id) @single_use = args[:single_use] if args.key?(:single_use) end end # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. class Type include Google::Apis::Core::Hashable # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. # Corresponds to the JSON property `arrayElementType` # @return [Google::Apis::SpannerV1::Type] attr_accessor :array_element_type # Required. The TypeCode for this type. # Corresponds to the JSON property `code` # @return [String] attr_accessor :code # `StructType` defines the fields of a STRUCT type. # Corresponds to the JSON property `structType` # @return [Google::Apis::SpannerV1::StructType] attr_accessor :struct_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @array_element_type = args[:array_element_type] if args.key?(:array_element_type) @code = args[:code] if args.key?(:code) @struct_type = args[:struct_type] if args.key?(:struct_type) end end # Metadata type for the operation returned by # UpdateDatabaseDdl. class UpdateDatabaseDdlMetadata include Google::Apis::Core::Hashable # Reports the commit timestamps of all statements that have # succeeded so far, where `commit_timestamps[i]` is the commit # timestamp for the statement `statements[i]`. # Corresponds to the JSON property `commitTimestamps` # @return [Array] attr_accessor :commit_timestamps # The database being modified. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # For an update this list contains all the statements. For an # individual statement, this list contains only that statement. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @commit_timestamps = args[:commit_timestamps] if args.key?(:commit_timestamps) @database = args[:database] if args.key?(:database) @statements = args[:statements] if args.key?(:statements) end end # Enqueues the given DDL statements to be applied, in order but not # necessarily all at once, to the database schema at some point (or # points) in the future. The server checks that the statements # are executable (syntactically valid, name tables that exist, etc.) # before enqueueing them, but they may still fail upon # later execution (e.g., if a statement from another batch of # statements is applied first and it conflicts in some way, or if # there is some data-related problem like a `NULL` value in a column to # which `NOT NULL` would be added). If a statement fails, all # subsequent statements in the batch are automatically cancelled. # Each batch of statements is assigned a name which can be used with # the Operations API to monitor # progress. See the # operation_id field for more # details. class UpdateDatabaseDdlRequest include Google::Apis::Core::Hashable # If empty, the new update request is assigned an # automatically-generated operation ID. Otherwise, `operation_id` # is used to construct the name of the resulting # Operation. # Specifying an explicit operation ID simplifies determining # whether the statements were executed in the event that the # UpdateDatabaseDdl call is replayed, # or the return value is otherwise lost: the database and # `operation_id` fields can be combined to form the # name of the resulting # longrunning.Operation: `/operations/`. # `operation_id` should be unique within the database, and must be # a valid identifier: `a-z*`. Note that # automatically-generated operation IDs always begin with an # underscore. If the named operation already exists, # UpdateDatabaseDdl returns # `ALREADY_EXISTS`. # Corresponds to the JSON property `operationId` # @return [String] attr_accessor :operation_id # Required. DDL statements to be applied to the database. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @operation_id = args[:operation_id] if args.key?(:operation_id) @statements = args[:statements] if args.key?(:statements) end end # Metadata type for the operation returned by # UpdateInstance. class UpdateInstanceMetadata include Google::Apis::Core::Hashable # The time at which this operation was cancelled. If set, this operation is # in the process of undoing itself (which is guaranteed to succeed) and # cannot be cancelled again. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The time at which this operation failed or was completed successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # The time at which UpdateInstance # request was received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @end_time = args[:end_time] if args.key?(:end_time) @instance = args[:instance] if args.key?(:instance) @start_time = args[:start_time] if args.key?(:start_time) end end # The request for UpdateInstance. class UpdateInstanceRequest include Google::Apis::Core::Hashable # Required. A mask specifying which fields in Instance should be updated. # The field mask must always be specified; this prevents any future fields in # Instance from being erased accidentally by clients that do not know # about them. # Corresponds to the JSON property `fieldMask` # @return [String] attr_accessor :field_mask # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @field_mask = args[:field_mask] if args.key?(:field_mask) @instance = args[:instance] if args.key?(:instance) end end # Arguments to insert, update, insert_or_update, and # replace operations. class Write include Google::Apis::Core::Hashable # The names of the columns in table to be written. # The list of columns must contain enough columns to allow # Cloud Spanner to derive values for all primary key columns in the # row(s) to be modified. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # Required. The table whose rows will be written. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # The values to be written. `values` can contain more than one # list of values. If it does, then multiple rows are written, one # for each entry in `values`. Each list in `values` must have # exactly as many entries as there are entries in columns # above. Sending multiple lists is equivalent to sending multiple # `Mutation`s, each containing one `values` entry and repeating # table and columns. Individual values in each list are # encoded as described here. # Corresponds to the JSON property `values` # @return [Array>] attr_accessor :values def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @table = args[:table] if args.key?(:table) @values = args[:values] if args.key?(:values) end end end end end