# Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. require 'date' require 'google/apis/core/base_service' require 'google/apis/core/json_representation' require 'google/apis/core/hashable' require 'google/apis/errors' module Google module Apis module SpannerV1 # A backup of a Cloud Spanner database. class Backup include Google::Apis::Core::Hashable # Output only. The backup will contain an externally consistent copy of the # database at the timestamp specified by `create_time`. `create_time` is # approximately the time the CreateBackup request is received. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Required for the CreateBackup operation. Name of the database from which this # backup was created. This needs to be in the same instance as the backup. # Values are of the form `projects//instances//databases/`. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # Required for the CreateBackup operation. The expiration time of the backup, # with microseconds granularity that must be at least 6 hours and at most 366 # days from the time the CreateBackup request is processed. Once the ` # expire_time` has passed, the backup is eligible to be automatically deleted by # Cloud Spanner to free the resources used by the backup. # Corresponds to the JSON property `expireTime` # @return [String] attr_accessor :expire_time # Output only for the CreateBackup operation. Required for the UpdateBackup # operation. A globally unique identifier for the backup which cannot be changed. # Values are of the form `projects//instances//backups/a-z*[a-z0-9]` The final # segment of the name must be between 2 and 60 characters in length. The backup # is stored in the location(s) specified in the instance configuration of the # instance containing the backup, identified by the prefix of the backup name of # the form `projects//instances/`. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Output only. The names of the restored databases that reference the backup. # The database names are of the form `projects//instances//databases/`. # Referencing databases may exist in different instances. The existence of any # referencing database prevents the backup from being deleted. When a restored # database from the backup enters the `READY` state, the reference to the backup # is removed. # Corresponds to the JSON property `referencingDatabases` # @return [Array] attr_accessor :referencing_databases # Output only. Size of the backup in bytes. # Corresponds to the JSON property `sizeBytes` # @return [Fixnum] attr_accessor :size_bytes # Output only. The current state of the backup. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_time = args[:create_time] if args.key?(:create_time) @database = args[:database] if args.key?(:database) @expire_time = args[:expire_time] if args.key?(:expire_time) @name = args[:name] if args.key?(:name) @referencing_databases = args[:referencing_databases] if args.key?(:referencing_databases) @size_bytes = args[:size_bytes] if args.key?(:size_bytes) @state = args[:state] if args.key?(:state) end end # Information about a backup. class BackupInfo include Google::Apis::Core::Hashable # Name of the backup. # Corresponds to the JSON property `backup` # @return [String] attr_accessor :backup # The backup contains an externally consistent copy of `source_database` at the # timestamp specified by `create_time`. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Name of the database the backup was created from. # Corresponds to the JSON property `sourceDatabase` # @return [String] attr_accessor :source_database def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup = args[:backup] if args.key?(:backup) @create_time = args[:create_time] if args.key?(:create_time) @source_database = args[:source_database] if args.key?(:source_database) end end # The request for BatchCreateSessions. class BatchCreateSessionsRequest include Google::Apis::Core::Hashable # Required. The number of sessions to be created in this batch call. The API may # return fewer than the requested number of sessions. If a specific number of # sessions are desired, the client can make additional calls to # BatchCreateSessions (adjusting session_count as necessary). # Corresponds to the JSON property `sessionCount` # @return [Fixnum] attr_accessor :session_count # A session in the Cloud Spanner API. # Corresponds to the JSON property `sessionTemplate` # @return [Google::Apis::SpannerV1::Session] attr_accessor :session_template def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session_count = args[:session_count] if args.key?(:session_count) @session_template = args[:session_template] if args.key?(:session_template) end end # The response for BatchCreateSessions. class BatchCreateSessionsResponse include Google::Apis::Core::Hashable # The freshly created sessions. # Corresponds to the JSON property `session` # @return [Array] attr_accessor :session def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session = args[:session] if args.key?(:session) end end # The request for BeginTransaction. class BeginTransactionRequest include Google::Apis::Core::Hashable # # Transactions Each session can have at most one active transaction at a time ( # note that standalone reads and queries use a transaction internally and do # count towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. It # is not necessary to create a new session for each transaction. # Transaction # Modes Cloud Spanner supports three transaction modes: 1. Locking read-write. # This type of transaction is the only way to write data into Cloud Spanner. # These transactions rely on pessimistic locking and, if necessary, two-phase # commit. Locking read-write transactions may abort, requiring the application # to retry. 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow writes. Snapshot read- # only transactions can be configured to read at timestamps in the past. # Snapshot read-only transactions do not need to be committed. 3. Partitioned # DML. This type of transaction is used to execute a single Partitioned DML # statement. Partitioned DML partitions the key space and runs the DML statement # over each partition in parallel using separate, internal transactions that # commit independently. Partitioned DML transactions do not need to be committed. # For transactions that only read, snapshot read-only transactions provide # simpler semantics and are almost always faster. In particular, read-only # transactions do not take locks, so they do not conflict with read-write # transactions. As a consequence of not taking locks, they also do not abort, so # retry loops are not needed. Transactions may only read/write data in a single # database. They may, however, read/write data in different tables within that # database. ## Locking Read-Write Transactions Locking transactions may be used # to atomically read-modify-write data anywhere in a database. This type of # transaction is externally consistent. Clients should attempt to minimize the # amount of time a transaction is active. Faster transactions commit with higher # probability and cause less contention. Cloud Spanner attempts to keep read # locks active as long as the transaction continues to do reads, and the # transaction has not been terminated by Commit or Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a transaction's # locks and abort it. Conceptually, a read-write transaction consists of zero or # more reads or SQL statements followed by Commit. At any time before Commit, # the client can send a Rollback request to abort the transaction. ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired are # still valid at commit time, and it is able to acquire write locks for all # writes. Cloud Spanner can abort the transaction for any reason. If a commit # attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has # not modified any user data in Cloud Spanner. Unless the transaction commits, # Cloud Spanner makes no guarantees about how long the transaction's locks were # held for. It is an error to use Cloud Spanner locks for any sort of mutual # exclusion other than between Cloud Spanner transactions themselves. ### # Retrying Aborted Transactions When a transaction aborts, the application can # choose to retry the whole transaction again. To maximize the chances of # successfully committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock priority # increases with each consecutive abort, meaning that each attempt has a # slightly better chance of success than the previous. Under some circumstances ( # e.g., many transactions attempting to modify the same row(s)), a transaction # can abort many times in a short period before successfully committing. Thus, # it is not a good idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent retrying. ## # # Idle Transactions A transaction is considered idle if it has no outstanding # reads or SQL queries and has not started a read or SQL query within the last # 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don' # t hold on to locks indefinitely. In that case, the commit will fail with error # `ABORTED`. If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from # becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only # transactions provides a simpler method than locking read-write transactions # for doing several consistent reads. However, this type of transaction does not # support writes. Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. # Since they do not acquire locks, they do not block concurrent read-write # transactions. Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read timestamp is # garbage collected; however, the default garbage collection policy is generous # enough that most applications do not need to worry about this in practice. # Snapshot read-only transactions do not need to call Commit or Rollback (and in # fact are not permitted to do so). To execute a snapshot transaction, the # client specifies a timestamp bound, which tells Cloud Spanner how to choose a # read timestamp. The types of timestamp bound are: - Strong (the default). - # Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read # is geographically distributed, stale read-only transactions can execute more # quickly than strong or read-write transaction, because they are able to # execute far from the leader replica. Each type of timestamp bound is discussed # in detail below. ### Strong Strong reads are guaranteed to see the effects of # all transactions that have committed before the start of the read. Furthermore, # all rows yielded by a single read are consistent with each other -- if any # part of the read observes a transaction, all parts of the read see the # transaction. Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are concurrent writes. # If consistency across reads is required, the reads should be executed within a # transaction or at an exact read timestamp. See TransactionOptions.ReadOnly. # strong. ### Exact Staleness These timestamp bounds execute reads at a user- # specified timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe modifications done by # all transactions with a commit timestamp <= the read timestamp, and observe # none of the modifications done by transactions with a larger commit timestamp. # They will block until all conflicting transactions that may be assigned commit # timestamps <= the read timestamp have finished. The timestamp can either be # expressed as an absolute Cloud Spanner commit timestamp or a staleness # relative to the current time. These modes do not require a "negotiation phase" # to pick a timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, boundedly # stale reads usually return fresher results. See TransactionOptions.ReadOnly. # read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded # Staleness Bounded staleness modes allow Cloud Spanner to pick the read # timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses # the newest timestamp within the staleness bound that allows execution of the # reads at the closest available replica without blocking. All rows yielded are # consistent with each other -- if any part of the read observes a transaction, # all parts of the read see the transaction. Boundedly stale reads are not # repeatable: two stale reads, even if they use the same staleness bound, can # execute at different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase negotiates a # timestamp among all replicas needed to serve the read. In the second phase, # reads are executed at the negotiated timestamp. As a result of the two phase # execution, bounded staleness reads are usually a little slower than comparable # exact staleness reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. Because the # timestamp negotiation requires up-front knowledge of which rows will be read, # it can only be used with single-use read-only transactions. See # TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly. # min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud # Spanner continuously garbage collects deleted and overwritten data in the # background to reclaim storage space. This process is known as "version GC". By # default, version GC reclaims versions after they are one hour old. Because of # this, Cloud Spanner cannot perform reads at read timestamps more than one hour # in the past. This restriction also applies to in-progress reads and/or SQL # queries whose timestamp become too old while executing. Reads and SQL queries # with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ## # Partitioned DML Transactions Partitioned DML transactions are used to execute # DML statements with a different execution strategy that provides different, # and often better, scalability properties for large, table-wide operations than # DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP # workload, should prefer using ReadWrite transactions. Partitioned DML # partitions the keyspace and runs the DML statement on each partition in # separate, internal transactions. These transactions commit automatically when # complete, and run independently from one another. To reduce lock contention, # this execution strategy only acquires read locks on rows that match the WHERE # clause of the statement. Additionally, the smaller per-partition transactions # hold locks for less time. That said, Partitioned DML is not a drop-in # replacement for standard DML used in ReadWrite transactions. - The DML # statement must be fully-partitionable. Specifically, the statement must be # expressible as the union of many statements which each access only a single # row of the table. - The statement is not applied atomically to all rows of the # table. Rather, the statement is applied atomically to partitions of the table, # in independent transactions. Secondary index rows are updated atomically with # the base table rows. - Partitioned DML does not guarantee exactly-once # execution semantics against a partition. The statement will be applied at # least once to each partition. It is strongly recommended that the DML # statement should be idempotent to avoid unexpected results. For instance, it # is potentially dangerous to run a statement such as `UPDATE table SET column = # column + 1` as it could be run multiple times against some rows. - The # partitions are committed automatically - there is no support for Commit or # Rollback. If the call returns an error, or if the client issuing the # ExecuteSql call dies, it is possible that some rows had the statement executed # on them successfully. It is also possible that statement was never executed # against other rows. - Partitioned DML transactions may only contain the # execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - # If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the operation is # stopped at that point and an error is returned. It is possible that at this # point, some partitions have been committed (or even committed multiple times), # and other partitions have not been run at all. Given the above, Partitioned # DML is good fit for large, database-wide, operations that are idempotent, such # as deleting old rows from a very large table. # Corresponds to the JSON property `options` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :options def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @options = args[:options] if args.key?(:options) end end # Associates `members` with a `role`. class Binding include Google::Apis::Core::Hashable # A client-specified ID for this binding. Expected to be globally unique to # support the internal bindings-by-ID API. # Corresponds to the JSON property `bindingId` # @return [String] attr_accessor :binding_id # Represents a textual expression in the Common Expression Language (CEL) syntax. # CEL is a C-like expression language. The syntax and semantics of CEL are # documented at https://github.com/google/cel-spec. Example (Comparison): title: # "Summary size limit" description: "Determines if a summary is less than 100 # chars" expression: "document.summary.size() < 100" Example (Equality): title: " # Requestor is owner" description: "Determines if requestor is the document # owner" expression: "document.owner == request.auth.claims.email" Example ( # Logic): title: "Public documents" description: "Determine whether the document # should be publicly visible" expression: "document.type != 'private' && # document.type != 'internal'" Example (Data Manipulation): title: "Notification # string" description: "Create a notification string with a timestamp." # expression: "'New message received at ' + string(document.create_time)" The # exact variables and functions that may be referenced within an expression are # determined by the service that evaluates it. See the service documentation for # additional information. # Corresponds to the JSON property `condition` # @return [Google::Apis::SpannerV1::Expr] attr_accessor :condition # Specifies the identities requesting access for a Cloud Platform resource. ` # members` can have the following values: * `allUsers`: A special identifier # that represents anyone who is on the internet; with or without a Google # account. * `allAuthenticatedUsers`: A special identifier that represents # anyone who is authenticated with a Google account or a service account. * ` # user:`emailid``: An email address that represents a specific Google account. # For example, `alice@example.com` . * `serviceAccount:`emailid``: An email # address that represents a service account. For example, `my-other-app@appspot. # gserviceaccount.com`. * `group:`emailid``: An email address that represents a # Google group. For example, `admins@example.com`. * `deleted:user:`emailid`?uid= # `uniqueid``: An email address (plus unique identifier) representing a user # that has been recently deleted. For example, `alice@example.com?uid= # 123456789012345678901`. If the user is recovered, this value reverts to `user:` # emailid`` and the recovered user retains the role in the binding. * `deleted: # serviceAccount:`emailid`?uid=`uniqueid``: An email address (plus unique # identifier) representing a service account that has been recently deleted. For # example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. # If the service account is undeleted, this value reverts to `serviceAccount:` # emailid`` and the undeleted service account retains the role in the binding. * # `deleted:group:`emailid`?uid=`uniqueid``: An email address (plus unique # identifier) representing a Google group that has been recently deleted. For # example, `admins@example.com?uid=123456789012345678901`. If the group is # recovered, this value reverts to `group:`emailid`` and the recovered group # retains the role in the binding. * `domain:`domain``: The G Suite domain ( # primary) that represents all the users of that domain. For example, `google. # com` or `example.com`. # Corresponds to the JSON property `members` # @return [Array] attr_accessor :members # Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor` # , or `roles/owner`. # Corresponds to the JSON property `role` # @return [String] attr_accessor :role def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @binding_id = args[:binding_id] if args.key?(:binding_id) @condition = args[:condition] if args.key?(:condition) @members = args[:members] if args.key?(:members) @role = args[:role] if args.key?(:role) end end # Metadata associated with a parent-child relationship appearing in a PlanNode. class ChildLink include Google::Apis::Core::Hashable # The node to which the link points. # Corresponds to the JSON property `childIndex` # @return [Fixnum] attr_accessor :child_index # The type of the link. For example, in Hash Joins this could be used to # distinguish between the build child and the probe child, or in the case of the # child being an output variable, to represent the tag associated with the # output variable. # Corresponds to the JSON property `type` # @return [String] attr_accessor :type # Only present if the child node is SCALAR and corresponds to an output variable # of the parent node. The field carries the name of the output variable. For # example, a `TableScan` operator that reads rows from a table will have child # links to the `SCALAR` nodes representing the output variables created for each # column that is read by the operator. The corresponding `variable` fields will # be set to the variable names assigned to the columns. # Corresponds to the JSON property `variable` # @return [String] attr_accessor :variable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @child_index = args[:child_index] if args.key?(:child_index) @type = args[:type] if args.key?(:type) @variable = args[:variable] if args.key?(:variable) end end # The request for Commit. class CommitRequest include Google::Apis::Core::Hashable # The mutations to be executed when this transaction commits. All mutations are # applied atomically, in the order they appear in this list. # Corresponds to the JSON property `mutations` # @return [Array] attr_accessor :mutations # # Transactions Each session can have at most one active transaction at a time ( # note that standalone reads and queries use a transaction internally and do # count towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. It # is not necessary to create a new session for each transaction. # Transaction # Modes Cloud Spanner supports three transaction modes: 1. Locking read-write. # This type of transaction is the only way to write data into Cloud Spanner. # These transactions rely on pessimistic locking and, if necessary, two-phase # commit. Locking read-write transactions may abort, requiring the application # to retry. 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow writes. Snapshot read- # only transactions can be configured to read at timestamps in the past. # Snapshot read-only transactions do not need to be committed. 3. Partitioned # DML. This type of transaction is used to execute a single Partitioned DML # statement. Partitioned DML partitions the key space and runs the DML statement # over each partition in parallel using separate, internal transactions that # commit independently. Partitioned DML transactions do not need to be committed. # For transactions that only read, snapshot read-only transactions provide # simpler semantics and are almost always faster. In particular, read-only # transactions do not take locks, so they do not conflict with read-write # transactions. As a consequence of not taking locks, they also do not abort, so # retry loops are not needed. Transactions may only read/write data in a single # database. They may, however, read/write data in different tables within that # database. ## Locking Read-Write Transactions Locking transactions may be used # to atomically read-modify-write data anywhere in a database. This type of # transaction is externally consistent. Clients should attempt to minimize the # amount of time a transaction is active. Faster transactions commit with higher # probability and cause less contention. Cloud Spanner attempts to keep read # locks active as long as the transaction continues to do reads, and the # transaction has not been terminated by Commit or Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a transaction's # locks and abort it. Conceptually, a read-write transaction consists of zero or # more reads or SQL statements followed by Commit. At any time before Commit, # the client can send a Rollback request to abort the transaction. ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired are # still valid at commit time, and it is able to acquire write locks for all # writes. Cloud Spanner can abort the transaction for any reason. If a commit # attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has # not modified any user data in Cloud Spanner. Unless the transaction commits, # Cloud Spanner makes no guarantees about how long the transaction's locks were # held for. It is an error to use Cloud Spanner locks for any sort of mutual # exclusion other than between Cloud Spanner transactions themselves. ### # Retrying Aborted Transactions When a transaction aborts, the application can # choose to retry the whole transaction again. To maximize the chances of # successfully committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock priority # increases with each consecutive abort, meaning that each attempt has a # slightly better chance of success than the previous. Under some circumstances ( # e.g., many transactions attempting to modify the same row(s)), a transaction # can abort many times in a short period before successfully committing. Thus, # it is not a good idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent retrying. ## # # Idle Transactions A transaction is considered idle if it has no outstanding # reads or SQL queries and has not started a read or SQL query within the last # 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don' # t hold on to locks indefinitely. In that case, the commit will fail with error # `ABORTED`. If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from # becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only # transactions provides a simpler method than locking read-write transactions # for doing several consistent reads. However, this type of transaction does not # support writes. Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. # Since they do not acquire locks, they do not block concurrent read-write # transactions. Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read timestamp is # garbage collected; however, the default garbage collection policy is generous # enough that most applications do not need to worry about this in practice. # Snapshot read-only transactions do not need to call Commit or Rollback (and in # fact are not permitted to do so). To execute a snapshot transaction, the # client specifies a timestamp bound, which tells Cloud Spanner how to choose a # read timestamp. The types of timestamp bound are: - Strong (the default). - # Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read # is geographically distributed, stale read-only transactions can execute more # quickly than strong or read-write transaction, because they are able to # execute far from the leader replica. Each type of timestamp bound is discussed # in detail below. ### Strong Strong reads are guaranteed to see the effects of # all transactions that have committed before the start of the read. Furthermore, # all rows yielded by a single read are consistent with each other -- if any # part of the read observes a transaction, all parts of the read see the # transaction. Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are concurrent writes. # If consistency across reads is required, the reads should be executed within a # transaction or at an exact read timestamp. See TransactionOptions.ReadOnly. # strong. ### Exact Staleness These timestamp bounds execute reads at a user- # specified timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe modifications done by # all transactions with a commit timestamp <= the read timestamp, and observe # none of the modifications done by transactions with a larger commit timestamp. # They will block until all conflicting transactions that may be assigned commit # timestamps <= the read timestamp have finished. The timestamp can either be # expressed as an absolute Cloud Spanner commit timestamp or a staleness # relative to the current time. These modes do not require a "negotiation phase" # to pick a timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, boundedly # stale reads usually return fresher results. See TransactionOptions.ReadOnly. # read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded # Staleness Bounded staleness modes allow Cloud Spanner to pick the read # timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses # the newest timestamp within the staleness bound that allows execution of the # reads at the closest available replica without blocking. All rows yielded are # consistent with each other -- if any part of the read observes a transaction, # all parts of the read see the transaction. Boundedly stale reads are not # repeatable: two stale reads, even if they use the same staleness bound, can # execute at different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase negotiates a # timestamp among all replicas needed to serve the read. In the second phase, # reads are executed at the negotiated timestamp. As a result of the two phase # execution, bounded staleness reads are usually a little slower than comparable # exact staleness reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. Because the # timestamp negotiation requires up-front knowledge of which rows will be read, # it can only be used with single-use read-only transactions. See # TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly. # min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud # Spanner continuously garbage collects deleted and overwritten data in the # background to reclaim storage space. This process is known as "version GC". By # default, version GC reclaims versions after they are one hour old. Because of # this, Cloud Spanner cannot perform reads at read timestamps more than one hour # in the past. This restriction also applies to in-progress reads and/or SQL # queries whose timestamp become too old while executing. Reads and SQL queries # with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ## # Partitioned DML Transactions Partitioned DML transactions are used to execute # DML statements with a different execution strategy that provides different, # and often better, scalability properties for large, table-wide operations than # DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP # workload, should prefer using ReadWrite transactions. Partitioned DML # partitions the keyspace and runs the DML statement on each partition in # separate, internal transactions. These transactions commit automatically when # complete, and run independently from one another. To reduce lock contention, # this execution strategy only acquires read locks on rows that match the WHERE # clause of the statement. Additionally, the smaller per-partition transactions # hold locks for less time. That said, Partitioned DML is not a drop-in # replacement for standard DML used in ReadWrite transactions. - The DML # statement must be fully-partitionable. Specifically, the statement must be # expressible as the union of many statements which each access only a single # row of the table. - The statement is not applied atomically to all rows of the # table. Rather, the statement is applied atomically to partitions of the table, # in independent transactions. Secondary index rows are updated atomically with # the base table rows. - Partitioned DML does not guarantee exactly-once # execution semantics against a partition. The statement will be applied at # least once to each partition. It is strongly recommended that the DML # statement should be idempotent to avoid unexpected results. For instance, it # is potentially dangerous to run a statement such as `UPDATE table SET column = # column + 1` as it could be run multiple times against some rows. - The # partitions are committed automatically - there is no support for Commit or # Rollback. If the call returns an error, or if the client issuing the # ExecuteSql call dies, it is possible that some rows had the statement executed # on them successfully. It is also possible that statement was never executed # against other rows. - Partitioned DML transactions may only contain the # execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - # If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the operation is # stopped at that point and an error is returned. It is possible that at this # point, some partitions have been committed (or even committed multiple times), # and other partitions have not been run at all. Given the above, Partitioned # DML is good fit for large, database-wide, operations that are idempotent, such # as deleting old rows from a very large table. # Corresponds to the JSON property `singleUseTransaction` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :single_use_transaction # Commit a previously-started transaction. # Corresponds to the JSON property `transactionId` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :transaction_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @mutations = args[:mutations] if args.key?(:mutations) @single_use_transaction = args[:single_use_transaction] if args.key?(:single_use_transaction) @transaction_id = args[:transaction_id] if args.key?(:transaction_id) end end # The response for Commit. class CommitResponse include Google::Apis::Core::Hashable # The Cloud Spanner timestamp at which the transaction committed. # Corresponds to the JSON property `commitTimestamp` # @return [String] attr_accessor :commit_timestamp def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @commit_timestamp = args[:commit_timestamp] if args.key?(:commit_timestamp) end end # Metadata type for the operation returned by CreateBackup. class CreateBackupMetadata include Google::Apis::Core::Hashable # The time at which cancellation of this operation was received. Operations. # CancelOperation starts asynchronous cancellation on a long-running operation. # The server makes a best effort to cancel the operation, but success is not # guaranteed. Clients can use Operations.GetOperation or other methods to check # whether the cancellation succeeded or whether the operation completed despite # cancellation. On successful cancellation, the operation is not deleted; # instead, it becomes an operation with an Operation.error value with a google. # rpc.Status.code of 1, corresponding to `Code.CANCELLED`. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The name of the database the backup is created from. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # The name of the backup being created. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Encapsulates progress related information for a Cloud Spanner long running # operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @database = args[:database] if args.key?(:database) @name = args[:name] if args.key?(:name) @progress = args[:progress] if args.key?(:progress) end end # Metadata type for the operation returned by CreateDatabase. class CreateDatabaseMetadata include Google::Apis::Core::Hashable # The database being created. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @database = args[:database] if args.key?(:database) end end # The request for CreateDatabase. class CreateDatabaseRequest include Google::Apis::Core::Hashable # Required. A `CREATE DATABASE` statement, which specifies the ID of the new # database. The database ID must conform to the regular expression `a-z*[a-z0-9]` # and be between 2 and 30 characters in length. If the database ID is a # reserved word or if it contains a hyphen, the database ID must be enclosed in # backticks (`` ` ``). # Corresponds to the JSON property `createStatement` # @return [String] attr_accessor :create_statement # Optional. A list of DDL statements to run inside the newly created database. # Statements can create tables, indexes, etc. These statements execute # atomically with the creation of the database: if there is an error in any # statement, the database is not created. # Corresponds to the JSON property `extraStatements` # @return [Array] attr_accessor :extra_statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_statement = args[:create_statement] if args.key?(:create_statement) @extra_statements = args[:extra_statements] if args.key?(:extra_statements) end end # Metadata type for the operation returned by CreateInstance. class CreateInstanceMetadata include Google::Apis::Core::Hashable # The time at which this operation was cancelled. If set, this operation is in # the process of undoing itself (which is guaranteed to succeed) and cannot be # cancelled again. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The time at which this operation failed or was completed successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # The time at which the CreateInstance request was received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @end_time = args[:end_time] if args.key?(:end_time) @instance = args[:instance] if args.key?(:instance) @start_time = args[:start_time] if args.key?(:start_time) end end # The request for CreateInstance. class CreateInstanceRequest include Google::Apis::Core::Hashable # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # Required. The ID of the instance to create. Valid identifiers are of the form ` # a-z*[a-z0-9]` and must be between 2 and 64 characters in length. # Corresponds to the JSON property `instanceId` # @return [String] attr_accessor :instance_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance = args[:instance] if args.key?(:instance) @instance_id = args[:instance_id] if args.key?(:instance_id) end end # The request for CreateSession. class CreateSessionRequest include Google::Apis::Core::Hashable # A session in the Cloud Spanner API. # Corresponds to the JSON property `session` # @return [Google::Apis::SpannerV1::Session] attr_accessor :session def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @session = args[:session] if args.key?(:session) end end # A Cloud Spanner database. class Database include Google::Apis::Core::Hashable # Output only. If exists, the time at which the database creation started. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # Required. The name of the database. Values are of the form `projects// # instances//databases/`, where `` is as specified in the `CREATE DATABASE` # statement. This name can be passed to other API methods to identify the # database. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Information about the database restore. # Corresponds to the JSON property `restoreInfo` # @return [Google::Apis::SpannerV1::RestoreInfo] attr_accessor :restore_info # Output only. The current database state. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @create_time = args[:create_time] if args.key?(:create_time) @name = args[:name] if args.key?(:name) @restore_info = args[:restore_info] if args.key?(:restore_info) @state = args[:state] if args.key?(:state) end end # Arguments to delete operations. class Delete include Google::Apis::Core::Hashable # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the # keys are expected to be in the same table or index. The keys need not be # sorted in any particular way. If the same key is specified multiple times in # the set (for example if two ranges, two keys, or a key and a range overlap), # Cloud Spanner behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # Required. The table whose rows will be deleted. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @key_set = args[:key_set] if args.key?(:key_set) @table = args[:table] if args.key?(:table) end end # A generic empty message that you can re-use to avoid defining duplicated empty # messages in your APIs. A typical example is to use it as the request or the # response type of an API method. For instance: service Foo ` rpc Bar(google. # protobuf.Empty) returns (google.protobuf.Empty); ` The JSON representation for # `Empty` is empty JSON object ````. class Empty include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # The request for ExecuteBatchDml. class ExecuteBatchDmlRequest include Google::Apis::Core::Hashable # Required. A per-transaction sequence number used to identify this request. # This field makes each request idempotent such that if the request is received # multiple times, at most one will succeed. The sequence number must be # monotonically increasing within the transaction. If a request arrives for the # first time with an out-of-order sequence number, the transaction may be # aborted. Replays of previously handled requests will yield the same response # as the first execution. # Corresponds to the JSON property `seqno` # @return [Fixnum] attr_accessor :seqno # Required. The list of statements to execute in this batch. Statements are # executed serially, such that the effects of statement `i` are visible to # statement `i+1`. Each statement must be a DML statement. Execution stops at # the first failed statement; the remaining statements are not executed. Callers # must provide at least one statement. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @seqno = args[:seqno] if args.key?(:seqno) @statements = args[:statements] if args.key?(:statements) @transaction = args[:transaction] if args.key?(:transaction) end end # The response for ExecuteBatchDml. Contains a list of ResultSet messages, one # for each DML statement that has successfully executed, in the same order as # the statements in the request. If a statement fails, the status in the # response body identifies the cause of the failure. To check for DML statements # that failed, use the following approach: 1. Check the status in the response # message. The google.rpc.Code enum value `OK` indicates that all statements # were executed successfully. 2. If the status was not `OK`, check the number of # result sets in the response. If the response contains `N` ResultSet messages, # then statement `N+1` in the request failed. Example 1: * Request: 5 DML # statements, all executed successfully. * Response: 5 ResultSet messages, with # the status `OK`. Example 2: * Request: 5 DML statements. The third statement # has a syntax error. * Response: 2 ResultSet messages, and a syntax error (` # INVALID_ARGUMENT`) status. The number of ResultSet messages indicates that the # third statement failed, and the fourth and fifth statements were not executed. class ExecuteBatchDmlResponse include Google::Apis::Core::Hashable # One ResultSet for each statement in the request that ran successfully, in the # same order as the statements in the request. Each ResultSet does not contain # any rows. The ResultSetStats in each ResultSet contain the number of rows # modified by the statement. Only the first ResultSet in the response contains # valid ResultSetMetadata. # Corresponds to the JSON property `resultSets` # @return [Array] attr_accessor :result_sets # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `status` # @return [Google::Apis::SpannerV1::Status] attr_accessor :status def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @result_sets = args[:result_sets] if args.key?(:result_sets) @status = args[:status] if args.key?(:status) end end # The request for ExecuteSql and ExecuteStreamingSql. class ExecuteSqlRequest include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type from a # JSON value. For example, values of type `BYTES` and values of type `STRING` # both appear in params as JSON strings. In these cases, `param_types` can be # used to specify the exact SQL type for some or all of the SQL statement # parameters. See the definition of Type for more information about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the SQL string. A # parameter placeholder consists of the `@` character followed by the parameter # name (for example, `@firstName`). Parameter names must conform to the naming # requirements of identifiers as specified at https://cloud.google.com/spanner/ # docs/lexical#identifiers. Parameters can appear anywhere that a literal value # is expected. The same parameter name can be used more than once, for example: ` # "WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL # statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # If present, results will be restricted to the specified partition previously # created using PartitionQuery(). There must be an exact match for the values of # fields common to this message and the PartitionQueryRequest message used to # create this partition_token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token # Used to control the amount of debugging information returned in ResultSetStats. # If partition_token is set, query_mode can only be set to QueryMode.NORMAL. # Corresponds to the JSON property `queryMode` # @return [String] attr_accessor :query_mode # Query optimizer configuration. # Corresponds to the JSON property `queryOptions` # @return [Google::Apis::SpannerV1::QueryOptions] attr_accessor :query_options # If this request is resuming a previously interrupted SQL statement execution, ` # resume_token` should be copied from the last PartialResultSet yielded before # the interruption. Doing this enables the new SQL statement execution to resume # where the last one left off. The rest of the request parameters must exactly # match the request that yielded this token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # A per-transaction sequence number used to identify this request. This field # makes each request idempotent such that if the request is received multiple # times, at most one will succeed. The sequence number must be monotonically # increasing within the transaction. If a request arrives for the first time # with an out-of-order sequence number, the transaction may be aborted. Replays # of previously handled requests will yield the same response as the first # execution. Required for DML statements. Ignored for queries. # Corresponds to the JSON property `seqno` # @return [Fixnum] attr_accessor :seqno # Required. The SQL string. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @partition_token = args[:partition_token] if args.key?(:partition_token) @query_mode = args[:query_mode] if args.key?(:query_mode) @query_options = args[:query_options] if args.key?(:query_options) @resume_token = args[:resume_token] if args.key?(:resume_token) @seqno = args[:seqno] if args.key?(:seqno) @sql = args[:sql] if args.key?(:sql) @transaction = args[:transaction] if args.key?(:transaction) end end # Represents a textual expression in the Common Expression Language (CEL) syntax. # CEL is a C-like expression language. The syntax and semantics of CEL are # documented at https://github.com/google/cel-spec. Example (Comparison): title: # "Summary size limit" description: "Determines if a summary is less than 100 # chars" expression: "document.summary.size() < 100" Example (Equality): title: " # Requestor is owner" description: "Determines if requestor is the document # owner" expression: "document.owner == request.auth.claims.email" Example ( # Logic): title: "Public documents" description: "Determine whether the document # should be publicly visible" expression: "document.type != 'private' && # document.type != 'internal'" Example (Data Manipulation): title: "Notification # string" description: "Create a notification string with a timestamp." # expression: "'New message received at ' + string(document.create_time)" The # exact variables and functions that may be referenced within an expression are # determined by the service that evaluates it. See the service documentation for # additional information. class Expr include Google::Apis::Core::Hashable # Optional. Description of the expression. This is a longer text which describes # the expression, e.g. when hovered over it in a UI. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description # Textual representation of an expression in Common Expression Language syntax. # Corresponds to the JSON property `expression` # @return [String] attr_accessor :expression # Optional. String indicating the location of the expression for error reporting, # e.g. a file name and a position in the file. # Corresponds to the JSON property `location` # @return [String] attr_accessor :location # Optional. Title for the expression, i.e. a short string describing its purpose. # This can be used e.g. in UIs which allow to enter the expression. # Corresponds to the JSON property `title` # @return [String] attr_accessor :title def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @description = args[:description] if args.key?(:description) @expression = args[:expression] if args.key?(:expression) @location = args[:location] if args.key?(:location) @title = args[:title] if args.key?(:title) end end # Message representing a single field of a struct. class Field include Google::Apis::Core::Hashable # The name of the field. For reads, this is the column name. For SQL queries, it # is the column alias (e.g., `"Word"` in the query `"SELECT 'hello' AS Word"`), # or the column name (e.g., `"ColName"` in the query `"SELECT ColName FROM Table" # `). Some columns might have an empty name (e.g., !"SELECT UPPER(ColName)"`). # Note that a query result can contain multiple fields with the same name. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. # Corresponds to the JSON property `type` # @return [Google::Apis::SpannerV1::Type] attr_accessor :type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @type = args[:type] if args.key?(:type) end end # The response for GetDatabaseDdl. class GetDatabaseDdlResponse include Google::Apis::Core::Hashable # A list of formatted DDL statements defining the schema of the database # specified in the request. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @statements = args[:statements] if args.key?(:statements) end end # Request message for `GetIamPolicy` method. class GetIamPolicyRequest include Google::Apis::Core::Hashable # Encapsulates settings provided to GetIamPolicy. # Corresponds to the JSON property `options` # @return [Google::Apis::SpannerV1::GetPolicyOptions] attr_accessor :options def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @options = args[:options] if args.key?(:options) end end # Encapsulates settings provided to GetIamPolicy. class GetPolicyOptions include Google::Apis::Core::Hashable # Optional. The policy format version to be returned. Valid values are 0, 1, and # 3. Requests specifying an invalid value will be rejected. Requests for # policies with any conditional bindings must specify version 3. Policies # without any conditional bindings may specify any valid value or leave the # field unset. To learn which resources support conditions in their IAM policies, # see the [IAM documentation](https://cloud.google.com/iam/help/conditions/ # resource-policies). # Corresponds to the JSON property `requestedPolicyVersion` # @return [Fixnum] attr_accessor :requested_policy_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @requested_policy_version = args[:requested_policy_version] if args.key?(:requested_policy_version) end end # An isolated set of Cloud Spanner resources on which databases can be hosted. class Instance include Google::Apis::Core::Hashable # Required. The name of the instance's configuration. Values are of the form ` # projects//instanceConfigs/`. See also InstanceConfig and ListInstanceConfigs. # Corresponds to the JSON property `config` # @return [String] attr_accessor :config # Required. The descriptive name for this instance as it appears in UIs. Must be # unique per project and between 4 and 30 characters in length. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # Deprecated. This field is not populated. # Corresponds to the JSON property `endpointUris` # @return [Array] attr_accessor :endpoint_uris # Cloud Labels are a flexible and lightweight mechanism for organizing cloud # resources into groups that reflect a customer's organizational needs and # deployment strategies. Cloud Labels can be used to filter collections of # resources. They can be used to control how resource metrics are aggregated. # And they can be used as arguments to policy management rules (e.g. route, # firewall, load balancing, etc.). * Label keys must be between 1 and 63 # characters long and must conform to the following regular expression: `[a-z]([- # a-z0-9]*[a-z0-9])?`. * Label values must be between 0 and 63 characters long # and must conform to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. * # No more than 64 labels can be associated with a given resource. See https:// # goo.gl/xmQnxf for more information on and examples of labels. If you plan to # use labels in your own code, please note that additional characters may be # allowed in the future. And so you are advised to use an internal label # representation, such as JSON, which doesn't rely upon specific characters # being disallowed. For example, representing labels as the string: name + "_" + # value would prove problematic if we were to allow "_" in a future release. # Corresponds to the JSON property `labels` # @return [Hash] attr_accessor :labels # Required. A unique identifier for the instance, which cannot be changed after # the instance is created. Values are of the form `projects//instances/a-z*[a-z0- # 9]`. The final segment of the name must be between 2 and 64 characters in # length. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The number of nodes allocated to this instance. This may be zero in API # responses for instances that are not yet in state `READY`. See [the # documentation](https://cloud.google.com/spanner/docs/instances#node_count) for # more information about nodes. # Corresponds to the JSON property `nodeCount` # @return [Fixnum] attr_accessor :node_count # Output only. The current instance state. For CreateInstance, the state must be # either omitted or set to `CREATING`. For UpdateInstance, the state must be # either omitted or set to `READY`. # Corresponds to the JSON property `state` # @return [String] attr_accessor :state def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @config = args[:config] if args.key?(:config) @display_name = args[:display_name] if args.key?(:display_name) @endpoint_uris = args[:endpoint_uris] if args.key?(:endpoint_uris) @labels = args[:labels] if args.key?(:labels) @name = args[:name] if args.key?(:name) @node_count = args[:node_count] if args.key?(:node_count) @state = args[:state] if args.key?(:state) end end # A possible configuration for a Cloud Spanner instance. Configurations define # the geographic placement of nodes and their replication. class InstanceConfig include Google::Apis::Core::Hashable # The name of this instance configuration as it appears in UIs. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # A unique identifier for the instance configuration. Values are of the form ` # projects//instanceConfigs/a-z*` # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The geographic placement of nodes in this instance configuration and their # replication properties. # Corresponds to the JSON property `replicas` # @return [Array] attr_accessor :replicas def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @display_name = args[:display_name] if args.key?(:display_name) @name = args[:name] if args.key?(:name) @replicas = args[:replicas] if args.key?(:replicas) end end # KeyRange represents a range of rows in a table or index. A range has a start # key and an end key. These keys can be open or closed, indicating if the range # includes rows with that key. Keys are represented by lists, where the ith # value in the list corresponds to the ith component of the table or index # primary key. Individual values are encoded as described here. For example, # consider the following table definition: CREATE TABLE UserEvents ( UserName # STRING(MAX), EventDate STRING(10) ) PRIMARY KEY(UserName, EventDate); The # following keys name rows in this table: "Bob", "2014-09-23" Since the ` # UserEvents` table's `PRIMARY KEY` clause names two columns, each `UserEvents` # key has two elements; the first is the `UserName`, and the second is the ` # EventDate`. Key ranges with multiple components are interpreted # lexicographically by component using the table or index key's declared sort # order. For example, the following range returns all events for user `"Bob"` # that occurred in the year 2015: "start_closed": ["Bob", "2015-01-01"] " # end_closed": ["Bob", "2015-12-31"] Start and end keys can omit trailing key # components. This affects the inclusion and exclusion of rows that exactly # match the provided key components: if the key is closed, then rows that # exactly match the provided components are included; if the key is open, then # rows that exactly match are not included. For example, the following range # includes all events for `"Bob"` that occurred during and after the year 2000: " # start_closed": ["Bob", "2000-01-01"] "end_closed": ["Bob"] The next example # retrieves all events for `"Bob"`: "start_closed": ["Bob"] "end_closed": ["Bob"] # To retrieve events before the year 2000: "start_closed": ["Bob"] "end_open": [ # "Bob", "2000-01-01"] The following range includes all rows in the table: " # start_closed": [] "end_closed": [] This range returns all users whose ` # UserName` begins with any character from A to C: "start_closed": ["A"] " # end_open": ["D"] This range returns all users whose `UserName` begins with B: " # start_closed": ["B"] "end_open": ["C"] Key ranges honor column sort order. For # example, suppose a table is defined as follows: CREATE TABLE # DescendingSortedTable ` Key INT64, ... ) PRIMARY KEY(Key DESC); The following # range retrieves all rows with key values between 1 and 100 inclusive: " # start_closed": ["100"] "end_closed": ["1"] Note that 100 is passed as the # start, and 1 is passed as the end, because `Key` is a descending column in the # schema. class KeyRange include Google::Apis::Core::Hashable # If the end is closed, then the range includes all rows whose first `len( # end_closed)` key columns exactly match `end_closed`. # Corresponds to the JSON property `endClosed` # @return [Array] attr_accessor :end_closed # If the end is open, then the range excludes rows whose first `len(end_open)` # key columns exactly match `end_open`. # Corresponds to the JSON property `endOpen` # @return [Array] attr_accessor :end_open # If the start is closed, then the range includes all rows whose first `len( # start_closed)` key columns exactly match `start_closed`. # Corresponds to the JSON property `startClosed` # @return [Array] attr_accessor :start_closed # If the start is open, then the range excludes rows whose first `len(start_open) # ` key columns exactly match `start_open`. # Corresponds to the JSON property `startOpen` # @return [Array] attr_accessor :start_open def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @end_closed = args[:end_closed] if args.key?(:end_closed) @end_open = args[:end_open] if args.key?(:end_open) @start_closed = args[:start_closed] if args.key?(:start_closed) @start_open = args[:start_open] if args.key?(:start_open) end end # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the # keys are expected to be in the same table or index. The keys need not be # sorted in any particular way. If the same key is specified multiple times in # the set (for example if two ranges, two keys, or a key and a range overlap), # Cloud Spanner behaves as if the key were only specified once. class KeySet include Google::Apis::Core::Hashable # For convenience `all` can be set to `true` to indicate that this `KeySet` # matches all keys in the table or index. Note that any keys specified in `keys` # or `ranges` are only yielded once. # Corresponds to the JSON property `all` # @return [Boolean] attr_accessor :all alias_method :all?, :all # A list of specific keys. Entries in `keys` should have exactly as many # elements as there are columns in the primary or index key with which this ` # KeySet` is used. Individual key values are encoded as described here. # Corresponds to the JSON property `keys` # @return [Array>] attr_accessor :keys # A list of key ranges. See KeyRange for more information about key range # specifications. # Corresponds to the JSON property `ranges` # @return [Array] attr_accessor :ranges def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @all = args[:all] if args.key?(:all) @keys = args[:keys] if args.key?(:keys) @ranges = args[:ranges] if args.key?(:ranges) end end # The response for ListBackupOperations. class ListBackupOperationsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent ListBackupOperations call to # fetch more of the matching metadata. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of matching backup long-running operations. Each operation's name # will be prefixed by the backup's name and the operation's metadata will be of # type CreateBackupMetadata. Operations returned include those that are pending # or have completed/failed/canceled within the last 7 days. Operations returned # are ordered by `operation.metadata.value.progress.start_time` in descending # order starting from the most recently started operation. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListBackups. class ListBackupsResponse include Google::Apis::Core::Hashable # The list of matching backups. Backups returned are ordered by `create_time` in # descending order, starting from the most recent `create_time`. # Corresponds to the JSON property `backups` # @return [Array] attr_accessor :backups # `next_page_token` can be sent in a subsequent ListBackups call to fetch more # of the matching backups. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backups = args[:backups] if args.key?(:backups) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for ListDatabaseOperations. class ListDatabaseOperationsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent ListDatabaseOperations call to # fetch more of the matching metadata. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of matching database long-running operations. Each operation's name # will be prefixed by the database's name. The operation's metadata field type ` # metadata.type_url` describes the type of the metadata. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListDatabases. class ListDatabasesResponse include Google::Apis::Core::Hashable # Databases that matched the request. # Corresponds to the JSON property `databases` # @return [Array] attr_accessor :databases # `next_page_token` can be sent in a subsequent ListDatabases call to fetch more # of the matching databases. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @databases = args[:databases] if args.key?(:databases) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for ListInstanceConfigs. class ListInstanceConfigsResponse include Google::Apis::Core::Hashable # The list of requested instance configurations. # Corresponds to the JSON property `instanceConfigs` # @return [Array] attr_accessor :instance_configs # `next_page_token` can be sent in a subsequent ListInstanceConfigs call to # fetch more of the matching instance configurations. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instance_configs = args[:instance_configs] if args.key?(:instance_configs) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response for ListInstances. class ListInstancesResponse include Google::Apis::Core::Hashable # The list of requested instances. # Corresponds to the JSON property `instances` # @return [Array] attr_accessor :instances # `next_page_token` can be sent in a subsequent ListInstances call to fetch more # of the matching instances. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @instances = args[:instances] if args.key?(:instances) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) end end # The response message for Operations.ListOperations. class ListOperationsResponse include Google::Apis::Core::Hashable # The standard List next-page token. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # A list of operations that matches the specified filter in the request. # Corresponds to the JSON property `operations` # @return [Array] attr_accessor :operations def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @operations = args[:operations] if args.key?(:operations) end end # The response for ListSessions. class ListSessionsResponse include Google::Apis::Core::Hashable # `next_page_token` can be sent in a subsequent ListSessions call to fetch more # of the matching sessions. # Corresponds to the JSON property `nextPageToken` # @return [String] attr_accessor :next_page_token # The list of requested sessions. # Corresponds to the JSON property `sessions` # @return [Array] attr_accessor :sessions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @next_page_token = args[:next_page_token] if args.key?(:next_page_token) @sessions = args[:sessions] if args.key?(:sessions) end end # A modification to one or more Cloud Spanner rows. Mutations can be applied to # a Cloud Spanner database by sending them in a Commit call. class Mutation include Google::Apis::Core::Hashable # Arguments to delete operations. # Corresponds to the JSON property `delete` # @return [Google::Apis::SpannerV1::Delete] attr_accessor :delete # Arguments to insert, update, insert_or_update, and replace operations. # Corresponds to the JSON property `insert` # @return [Google::Apis::SpannerV1::Write] attr_accessor :insert # Arguments to insert, update, insert_or_update, and replace operations. # Corresponds to the JSON property `insertOrUpdate` # @return [Google::Apis::SpannerV1::Write] attr_accessor :insert_or_update # Arguments to insert, update, insert_or_update, and replace operations. # Corresponds to the JSON property `replace` # @return [Google::Apis::SpannerV1::Write] attr_accessor :replace # Arguments to insert, update, insert_or_update, and replace operations. # Corresponds to the JSON property `update` # @return [Google::Apis::SpannerV1::Write] attr_accessor :update def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @delete = args[:delete] if args.key?(:delete) @insert = args[:insert] if args.key?(:insert) @insert_or_update = args[:insert_or_update] if args.key?(:insert_or_update) @replace = args[:replace] if args.key?(:replace) @update = args[:update] if args.key?(:update) end end # This resource represents a long-running operation that is the result of a # network API call. class Operation include Google::Apis::Core::Hashable # If the value is `false`, it means the operation is still in progress. If `true` # , the operation is completed, and either `error` or `response` is available. # Corresponds to the JSON property `done` # @return [Boolean] attr_accessor :done alias_method :done?, :done # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). # Corresponds to the JSON property `error` # @return [Google::Apis::SpannerV1::Status] attr_accessor :error # Service-specific metadata associated with the operation. It typically contains # progress information and common metadata such as create time. Some services # might not provide such metadata. Any method that returns a long-running # operation should document the metadata type, if any. # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # The server-assigned name, which is only unique within the same service that # originally returns it. If you use the default HTTP mapping, the `name` should # be a resource name ending with `operations/`unique_id``. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # The normal response of the operation in case of success. If the original # method returns no data on success, such as `Delete`, the response is `google. # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, # the response should be the resource. For other methods, the response should # have the type `XxxResponse`, where `Xxx` is the original method name. For # example, if the original method name is `TakeSnapshot()`, the inferred # response type is `TakeSnapshotResponse`. # Corresponds to the JSON property `response` # @return [Hash] attr_accessor :response def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @done = args[:done] if args.key?(:done) @error = args[:error] if args.key?(:error) @metadata = args[:metadata] if args.key?(:metadata) @name = args[:name] if args.key?(:name) @response = args[:response] if args.key?(:response) end end # Encapsulates progress related information for a Cloud Spanner long running # operation. class OperationProgress include Google::Apis::Core::Hashable # If set, the time at which this operation failed or was completed successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # Percent completion of the operation. Values are between 0 and 100 inclusive. # Corresponds to the JSON property `progressPercent` # @return [Fixnum] attr_accessor :progress_percent # Time the request was received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @end_time = args[:end_time] if args.key?(:end_time) @progress_percent = args[:progress_percent] if args.key?(:progress_percent) @start_time = args[:start_time] if args.key?(:start_time) end end # Metadata type for the long-running operation used to track the progress of # optimizations performed on a newly restored database. This long-running # operation is automatically created by the system after the successful # completion of a database restore, and cannot be cancelled. class OptimizeRestoredDatabaseMetadata include Google::Apis::Core::Hashable # Name of the restored database being optimized. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Encapsulates progress related information for a Cloud Spanner long running # operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @name = args[:name] if args.key?(:name) @progress = args[:progress] if args.key?(:progress) end end # Partial results from a streaming read or SQL query. Streaming reads and SQL # queries better tolerate large result sets, large rows, and large values, but # are a little trickier to consume. class PartialResultSet include Google::Apis::Core::Hashable # If true, then the final value in values is chunked, and must be combined with # more values from subsequent `PartialResultSet`s to obtain a complete field # value. # Corresponds to the JSON property `chunkedValue` # @return [Boolean] attr_accessor :chunked_value alias_method :chunked_value?, :chunked_value # Metadata about a ResultSet or PartialResultSet. # Corresponds to the JSON property `metadata` # @return [Google::Apis::SpannerV1::ResultSetMetadata] attr_accessor :metadata # Streaming calls might be interrupted for a variety of reasons, such as TCP # connection loss. If this occurs, the stream of results can be resumed by re- # sending the original request and including `resume_token`. Note that executing # any other transaction in the same session invalidates the token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # Additional statistics about a ResultSet or PartialResultSet. # Corresponds to the JSON property `stats` # @return [Google::Apis::SpannerV1::ResultSetStats] attr_accessor :stats # A streamed result set consists of a stream of values, which might be split # into many `PartialResultSet` messages to accommodate large rows and/or large # values. Every N complete values defines a row, where N is equal to the number # of entries in metadata.row_type.fields. Most values are encoded based on type # as described here. It is possible that the last value in values is "chunked", # meaning that the rest of the value is sent in subsequent `PartialResultSet`(s). # This is denoted by the chunked_value field. Two or more chunked values can be # merged to form a complete value as follows: * `bool/number/null`: cannot be # chunked * `string`: concatenate the strings * `list`: concatenate the lists. # If the last element in a list is a `string`, `list`, or `object`, merge it # with the first element in the next list by applying these rules recursively. * # `object`: concatenate the (field name, field value) pairs. If a field name is # duplicated, then apply these rules recursively to merge the field values. Some # examples of merging: # Strings are concatenated. "foo", "bar" => "foobar" # # Lists of non-strings are concatenated. [2, 3], [4] => [2, 3, 4] # Lists are # concatenated, but the last and first elements are merged # because they are # strings. ["a", "b"], ["c", "d"] => ["a", "bc", "d"] # Lists are concatenated, # but the last and first elements are merged # because they are lists. # Recursively, the last and first elements # of the inner lists are merged # because they are strings. ["a", ["b", "c"]], [["d"], "e"] => ["a", ["b", "cd"], # "e"] # Non-overlapping object fields are combined. `"a": "1"`, `"b": "2"` => ` # "a": "1", "b": 2"` # Overlapping object fields are merged. `"a": "1"`, `"a": " # 2"` => `"a": "12"` # Examples of merging objects containing lists of strings. ` # "a": ["1"]`, `"a": ["2"]` => `"a": ["12"]` For a more complete example, # suppose a streaming SQL query is yielding a result set whose rows contain a # single string field. The following `PartialResultSet`s might be yielded: ` " # metadata": ` ... ` "values": ["Hello", "W"] "chunked_value": true " # resume_token": "Af65..." ` ` "values": ["orl"] "chunked_value": true " # resume_token": "Bqp2..." ` ` "values": ["d"] "resume_token": "Zx1B..." ` This # sequence of `PartialResultSet`s encodes two rows, one containing the field # value `"Hello"`, and a second containing the field value `"World" = "W" + "orl" # + "d"`. # Corresponds to the JSON property `values` # @return [Array] attr_accessor :values def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @chunked_value = args[:chunked_value] if args.key?(:chunked_value) @metadata = args[:metadata] if args.key?(:metadata) @resume_token = args[:resume_token] if args.key?(:resume_token) @stats = args[:stats] if args.key?(:stats) @values = args[:values] if args.key?(:values) end end # Information returned for each partition returned in a PartitionResponse. class Partition include Google::Apis::Core::Hashable # This token can be passed to Read, StreamingRead, ExecuteSql, or # ExecuteStreamingSql requests to restrict the results to those identified by # this partition token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partition_token = args[:partition_token] if args.key?(:partition_token) end end # Options for a PartitionQueryRequest and PartitionReadRequest. class PartitionOptions include Google::Apis::Core::Hashable # **Note:** This hint is currently ignored by PartitionQuery and PartitionRead # requests. The desired maximum number of partitions to return. For example, # this may be set to the number of workers available. The default for this # option is currently 10,000. The maximum value is currently 200,000. This is # only a hint. The actual number of partitions returned may be smaller or larger # than this maximum count request. # Corresponds to the JSON property `maxPartitions` # @return [Fixnum] attr_accessor :max_partitions # **Note:** This hint is currently ignored by PartitionQuery and PartitionRead # requests. The desired data size for each partition generated. The default for # this option is currently 1 GiB. This is only a hint. The actual size of each # partition may be smaller or larger than this size request. # Corresponds to the JSON property `partitionSizeBytes` # @return [Fixnum] attr_accessor :partition_size_bytes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @max_partitions = args[:max_partitions] if args.key?(:max_partitions) @partition_size_bytes = args[:partition_size_bytes] if args.key?(:partition_size_bytes) end end # The request for PartitionQuery class PartitionQueryRequest include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type from a # JSON value. For example, values of type `BYTES` and values of type `STRING` # both appear in params as JSON strings. In these cases, `param_types` can be # used to specify the exact SQL type for some or all of the SQL query parameters. # See the definition of Type for more information about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the SQL string. A # parameter placeholder consists of the `@` character followed by the parameter # name (for example, `@firstName`). Parameter names can contain letters, numbers, # and underscores. Parameters can appear anywhere that a literal value is # expected. The same parameter name can be used more than once, for example: `" # WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL # statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # Options for a PartitionQueryRequest and PartitionReadRequest. # Corresponds to the JSON property `partitionOptions` # @return [Google::Apis::SpannerV1::PartitionOptions] attr_accessor :partition_options # Required. The query request to generate partitions for. The request will fail # if the query is not root partitionable. The query plan of a root partitionable # query has a single distributed union operator. A distributed union operator # conceptually divides one or more tables into multiple splits, remotely # evaluates a subquery independently on each split, and then unions all results. # This must not contain DML commands, such as INSERT, UPDATE, or DELETE. Use # ExecuteStreamingSql with a PartitionedDml transaction for large, partition- # friendly DML operations. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @partition_options = args[:partition_options] if args.key?(:partition_options) @sql = args[:sql] if args.key?(:sql) @transaction = args[:transaction] if args.key?(:transaction) end end # The request for PartitionRead class PartitionReadRequest include Google::Apis::Core::Hashable # The columns of table to be returned for each row matching this request. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # If non-empty, the name of an index on table. This index is used instead of the # table primary key when interpreting key_set and sorting result rows. See # key_set for further information. # Corresponds to the JSON property `index` # @return [String] attr_accessor :index # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the # keys are expected to be in the same table or index. The keys need not be # sorted in any particular way. If the same key is specified multiple times in # the set (for example if two ranges, two keys, or a key and a range overlap), # Cloud Spanner behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # Options for a PartitionQueryRequest and PartitionReadRequest. # Corresponds to the JSON property `partitionOptions` # @return [Google::Apis::SpannerV1::PartitionOptions] attr_accessor :partition_options # Required. The name of the table in the database to be read. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @index = args[:index] if args.key?(:index) @key_set = args[:key_set] if args.key?(:key_set) @partition_options = args[:partition_options] if args.key?(:partition_options) @table = args[:table] if args.key?(:table) @transaction = args[:transaction] if args.key?(:transaction) end end # The response for PartitionQuery or PartitionRead class PartitionResponse include Google::Apis::Core::Hashable # Partitions created by this request. # Corresponds to the JSON property `partitions` # @return [Array] attr_accessor :partitions # A transaction. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::Transaction] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partitions = args[:partitions] if args.key?(:partitions) @transaction = args[:transaction] if args.key?(:transaction) end end # Message type to initiate a Partitioned DML transaction. class PartitionedDml include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # Node information for nodes appearing in a QueryPlan.plan_nodes. class PlanNode include Google::Apis::Core::Hashable # List of child node `index`es and their relationship to this parent. # Corresponds to the JSON property `childLinks` # @return [Array] attr_accessor :child_links # The display name for the node. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # The execution statistics associated with the node, contained in a group of key- # value pairs. Only present if the plan was returned as a result of a profile # query. For example, number of executions, number of rows/time per execution # etc. # Corresponds to the JSON property `executionStats` # @return [Hash] attr_accessor :execution_stats # The `PlanNode`'s index in node list. # Corresponds to the JSON property `index` # @return [Fixnum] attr_accessor :index # Used to determine the type of node. May be needed for visualizing different # kinds of nodes differently. For example, If the node is a SCALAR node, it will # have a condensed representation which can be used to directly embed a # description of the node in its parent. # Corresponds to the JSON property `kind` # @return [String] attr_accessor :kind # Attributes relevant to the node contained in a group of key-value pairs. For # example, a Parameter Reference node could have the following information in # its metadata: ` "parameter_reference": "param1", "parameter_type": "array" ` # Corresponds to the JSON property `metadata` # @return [Hash] attr_accessor :metadata # Condensed representation of a node and its subtree. Only present for `SCALAR` # PlanNode(s). # Corresponds to the JSON property `shortRepresentation` # @return [Google::Apis::SpannerV1::ShortRepresentation] attr_accessor :short_representation def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @child_links = args[:child_links] if args.key?(:child_links) @display_name = args[:display_name] if args.key?(:display_name) @execution_stats = args[:execution_stats] if args.key?(:execution_stats) @index = args[:index] if args.key?(:index) @kind = args[:kind] if args.key?(:kind) @metadata = args[:metadata] if args.key?(:metadata) @short_representation = args[:short_representation] if args.key?(:short_representation) end end # An Identity and Access Management (IAM) policy, which specifies access # controls for Google Cloud resources. A `Policy` is a collection of `bindings`. # A `binding` binds one or more `members` to a single `role`. Members can be # user accounts, service accounts, Google groups, and domains (such as G Suite). # A `role` is a named list of permissions; each `role` can be an IAM predefined # role or a user-created custom role. For some types of Google Cloud resources, # a `binding` can also specify a `condition`, which is a logical expression that # allows access to a resource only if the expression evaluates to `true`. A # condition can add constraints based on attributes of the request, the resource, # or both. To learn which resources support conditions in their IAM policies, # see the [IAM documentation](https://cloud.google.com/iam/help/conditions/ # resource-policies). **JSON example:** ` "bindings": [ ` "role": "roles/ # resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", " # group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@ # appspot.gserviceaccount.com" ] `, ` "role": "roles/resourcemanager. # organizationViewer", "members": [ "user:eve@example.com" ], "condition": ` " # title": "expirable access", "description": "Does not grant access after Sep # 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", ` # ` ], "etag": "BwWWja0YfJA=", "version": 3 ` **YAML example:** bindings: - # members: - user:mike@example.com - group:admins@example.com - domain:google. # com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/ # resourcemanager.organizationAdmin - members: - user:eve@example.com role: # roles/resourcemanager.organizationViewer condition: title: expirable access # description: Does not grant access after Sep 2020 expression: request.time < # timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a # description of IAM and its features, see the [IAM documentation](https://cloud. # google.com/iam/docs/). class Policy include Google::Apis::Core::Hashable # Associates a list of `members` to a `role`. Optionally, may specify a ` # condition` that determines how and when the `bindings` are applied. Each of # the `bindings` must contain at least one member. # Corresponds to the JSON property `bindings` # @return [Array] attr_accessor :bindings # `etag` is used for optimistic concurrency control as a way to help prevent # simultaneous updates of a policy from overwriting each other. It is strongly # suggested that systems make use of the `etag` in the read-modify-write cycle # to perform policy updates in order to avoid race conditions: An `etag` is # returned in the response to `getIamPolicy`, and systems are expected to put # that etag in the request to `setIamPolicy` to ensure that their change will be # applied to the same version of the policy. **Important:** If you use IAM # Conditions, you must include the `etag` field whenever you call `setIamPolicy`. # If you omit this field, then IAM allows you to overwrite a version `3` policy # with a version `1` policy, and all of the conditions in the version `3` policy # are lost. # Corresponds to the JSON property `etag` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :etag # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. # Requests that specify an invalid value are rejected. Any operation that # affects conditional role bindings must specify version `3`. This requirement # applies to the following operations: * Getting a policy that includes a # conditional role binding * Adding a conditional role binding to a policy * # Changing a conditional role binding in a policy * Removing any role binding, # with or without a condition, from a policy that includes conditions ** # Important:** If you use IAM Conditions, you must include the `etag` field # whenever you call `setIamPolicy`. If you omit this field, then IAM allows you # to overwrite a version `3` policy with a version `1` policy, and all of the # conditions in the version `3` policy are lost. If a policy does not include # any conditions, operations on that policy may specify any valid version or # leave the field unset. To learn which resources support conditions in their # IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/ # conditions/resource-policies). # Corresponds to the JSON property `version` # @return [Fixnum] attr_accessor :version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @bindings = args[:bindings] if args.key?(:bindings) @etag = args[:etag] if args.key?(:etag) @version = args[:version] if args.key?(:version) end end # Query optimizer configuration. class QueryOptions include Google::Apis::Core::Hashable # An option to control the selection of optimizer version. This parameter allows # individual queries to pick different query optimizer versions. Specifying " # latest" as a value instructs Cloud Spanner to use the latest supported query # optimizer version. If not specified, Cloud Spanner uses optimizer version set # at the database level options. Any other positive integer (from the list of # supported optimizer versions) overrides the default optimizer version for # query execution. The list of supported optimizer versions can be queried from # SPANNER_SYS.SUPPORTED_OPTIMIZER_VERSIONS. Executing a SQL statement with an # invalid optimizer version will fail with a syntax error (`INVALID_ARGUMENT`) # status. See https://cloud.google.com/spanner/docs/query-optimizer/manage-query- # optimizer for more information on managing the query optimizer. The ` # optimizer_version` statement hint has precedence over this setting. # Corresponds to the JSON property `optimizerVersion` # @return [String] attr_accessor :optimizer_version def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @optimizer_version = args[:optimizer_version] if args.key?(:optimizer_version) end end # Contains an ordered list of nodes appearing in the query plan. class QueryPlan include Google::Apis::Core::Hashable # The nodes in the query plan. Plan nodes are returned in pre-order starting # with the plan root. Each PlanNode's `id` corresponds to its index in ` # plan_nodes`. # Corresponds to the JSON property `planNodes` # @return [Array] attr_accessor :plan_nodes def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @plan_nodes = args[:plan_nodes] if args.key?(:plan_nodes) end end # Message type to initiate a read-only transaction. class ReadOnly include Google::Apis::Core::Hashable # Executes all reads at a timestamp that is `exact_staleness` old. The timestamp # is chosen soon after the read is started. Guarantees that all writes that have # committed more than the specified number of seconds ago are visible. Because # Cloud Spanner chooses the exact timestamp, this mode works even if the client' # s local clock is substantially skewed from Cloud Spanner commit timestamps. # Useful for reading at nearby replicas without the distributed timestamp # negotiation overhead of `max_staleness`. # Corresponds to the JSON property `exactStaleness` # @return [String] attr_accessor :exact_staleness # Read data at a timestamp >= `NOW - max_staleness` seconds. Guarantees that all # writes that have committed more than the specified number of seconds ago are # visible. Because Cloud Spanner chooses the exact timestamp, this mode works # even if the client's local clock is substantially skewed from Cloud Spanner # commit timestamps. Useful for reading the freshest data available at a nearby # replica, while bounding the possible staleness if the local replica has fallen # behind. Note that this option can only be used in single-use transactions. # Corresponds to the JSON property `maxStaleness` # @return [String] attr_accessor :max_staleness # Executes all reads at a timestamp >= `min_read_timestamp`. This is useful for # requesting fresher data than some previous read, or data that is fresh enough # to observe the effects of some previously committed transaction whose # timestamp is known. Note that this option can only be used in single-use # transactions. A timestamp in RFC3339 UTC \"Zulu\" format, accurate to # nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `minReadTimestamp` # @return [String] attr_accessor :min_read_timestamp # Executes all reads at the given timestamp. Unlike other modes, reads at a # specific timestamp are repeatable; the same read at the same timestamp always # returns the same data. If the timestamp is in the future, the read will block # until the specified timestamp, modulo the read's deadline. Useful for large # scale consistent reads such as mapreduces, or for coordinating many reads # against a consistent snapshot of the data. A timestamp in RFC3339 UTC \"Zulu\" # format, accurate to nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `readTimestamp` # @return [String] attr_accessor :read_timestamp # If true, the Cloud Spanner-selected read timestamp is included in the # Transaction message that describes the transaction. # Corresponds to the JSON property `returnReadTimestamp` # @return [Boolean] attr_accessor :return_read_timestamp alias_method :return_read_timestamp?, :return_read_timestamp # Read at a timestamp where all previously committed transactions are visible. # Corresponds to the JSON property `strong` # @return [Boolean] attr_accessor :strong alias_method :strong?, :strong def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @exact_staleness = args[:exact_staleness] if args.key?(:exact_staleness) @max_staleness = args[:max_staleness] if args.key?(:max_staleness) @min_read_timestamp = args[:min_read_timestamp] if args.key?(:min_read_timestamp) @read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp) @return_read_timestamp = args[:return_read_timestamp] if args.key?(:return_read_timestamp) @strong = args[:strong] if args.key?(:strong) end end # The request for Read and StreamingRead. class ReadRequest include Google::Apis::Core::Hashable # Required. The columns of table to be returned for each row matching this # request. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # If non-empty, the name of an index on table. This index is used instead of the # table primary key when interpreting key_set and sorting result rows. See # key_set for further information. # Corresponds to the JSON property `index` # @return [String] attr_accessor :index # `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the # keys are expected to be in the same table or index. The keys need not be # sorted in any particular way. If the same key is specified multiple times in # the set (for example if two ranges, two keys, or a key and a range overlap), # Cloud Spanner behaves as if the key were only specified once. # Corresponds to the JSON property `keySet` # @return [Google::Apis::SpannerV1::KeySet] attr_accessor :key_set # If greater than zero, only the first `limit` rows are yielded. If `limit` is # zero, the default is no limit. A limit cannot be specified if `partition_token` # is set. # Corresponds to the JSON property `limit` # @return [Fixnum] attr_accessor :limit # If present, results will be restricted to the specified partition previously # created using PartitionRead(). There must be an exact match for the values of # fields common to this message and the PartitionReadRequest message used to # create this partition_token. # Corresponds to the JSON property `partitionToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :partition_token # If this request is resuming a previously interrupted read, `resume_token` # should be copied from the last PartialResultSet yielded before the # interruption. Doing this enables the new read to resume where the last read # left off. The rest of the request parameters must exactly match the request # that yielded this token. # Corresponds to the JSON property `resumeToken` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :resume_token # Required. The name of the table in the database to be read. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::TransactionSelector] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @index = args[:index] if args.key?(:index) @key_set = args[:key_set] if args.key?(:key_set) @limit = args[:limit] if args.key?(:limit) @partition_token = args[:partition_token] if args.key?(:partition_token) @resume_token = args[:resume_token] if args.key?(:resume_token) @table = args[:table] if args.key?(:table) @transaction = args[:transaction] if args.key?(:transaction) end end # Message type to initiate a read-write transaction. Currently this transaction # type has no options. class ReadWrite include Google::Apis::Core::Hashable def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) end end # class ReplicaInfo include Google::Apis::Core::Hashable # If true, this location is designated as the default leader location where # leader replicas are placed. See the [region types documentation](https://cloud. # google.com/spanner/docs/instances#region_types) for more details. # Corresponds to the JSON property `defaultLeaderLocation` # @return [Boolean] attr_accessor :default_leader_location alias_method :default_leader_location?, :default_leader_location # The location of the serving resources, e.g. "us-central1". # Corresponds to the JSON property `location` # @return [String] attr_accessor :location # The type of replica. # Corresponds to the JSON property `type` # @return [String] attr_accessor :type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @default_leader_location = args[:default_leader_location] if args.key?(:default_leader_location) @location = args[:location] if args.key?(:location) @type = args[:type] if args.key?(:type) end end # Metadata type for the long-running operation returned by RestoreDatabase. class RestoreDatabaseMetadata include Google::Apis::Core::Hashable # Information about a backup. # Corresponds to the JSON property `backupInfo` # @return [Google::Apis::SpannerV1::BackupInfo] attr_accessor :backup_info # The time at which cancellation of this operation was received. Operations. # CancelOperation starts asynchronous cancellation on a long-running operation. # The server makes a best effort to cancel the operation, but success is not # guaranteed. Clients can use Operations.GetOperation or other methods to check # whether the cancellation succeeded or whether the operation completed despite # cancellation. On successful cancellation, the operation is not deleted; # instead, it becomes an operation with an Operation.error value with a google. # rpc.Status.code of 1, corresponding to `Code.CANCELLED`. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # Name of the database being created and restored to. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # If exists, the name of the long-running operation that will be used to track # the post-restore optimization process to optimize the performance of the # restored database, and remove the dependency on the restore source. The name # is of the form `projects//instances//databases//operations/` where the is the # name of database being created and restored to. The metadata type of the long- # running operation is OptimizeRestoredDatabaseMetadata. This long-running # operation will be automatically created by the system after the # RestoreDatabase long-running operation completes successfully. This operation # will not be created if the restore was not successful. # Corresponds to the JSON property `optimizeDatabaseOperationName` # @return [String] attr_accessor :optimize_database_operation_name # Encapsulates progress related information for a Cloud Spanner long running # operation. # Corresponds to the JSON property `progress` # @return [Google::Apis::SpannerV1::OperationProgress] attr_accessor :progress # The type of the restore source. # Corresponds to the JSON property `sourceType` # @return [String] attr_accessor :source_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup_info = args[:backup_info] if args.key?(:backup_info) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @name = args[:name] if args.key?(:name) @optimize_database_operation_name = args[:optimize_database_operation_name] if args.key?(:optimize_database_operation_name) @progress = args[:progress] if args.key?(:progress) @source_type = args[:source_type] if args.key?(:source_type) end end # The request for RestoreDatabase. class RestoreDatabaseRequest include Google::Apis::Core::Hashable # Name of the backup from which to restore. Values are of the form `projects// # instances//backups/`. # Corresponds to the JSON property `backup` # @return [String] attr_accessor :backup # Required. The id of the database to create and restore to. This database must # not already exist. The `database_id` appended to `parent` forms the full # database name of the form `projects//instances//databases/`. # Corresponds to the JSON property `databaseId` # @return [String] attr_accessor :database_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup = args[:backup] if args.key?(:backup) @database_id = args[:database_id] if args.key?(:database_id) end end # Information about the database restore. class RestoreInfo include Google::Apis::Core::Hashable # Information about a backup. # Corresponds to the JSON property `backupInfo` # @return [Google::Apis::SpannerV1::BackupInfo] attr_accessor :backup_info # The type of the restore source. # Corresponds to the JSON property `sourceType` # @return [String] attr_accessor :source_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @backup_info = args[:backup_info] if args.key?(:backup_info) @source_type = args[:source_type] if args.key?(:source_type) end end # Results from Read or ExecuteSql. class ResultSet include Google::Apis::Core::Hashable # Metadata about a ResultSet or PartialResultSet. # Corresponds to the JSON property `metadata` # @return [Google::Apis::SpannerV1::ResultSetMetadata] attr_accessor :metadata # Each element in `rows` is a row whose format is defined by metadata.row_type. # The ith element in each row matches the ith field in metadata.row_type. # Elements are encoded based on type as described here. # Corresponds to the JSON property `rows` # @return [Array>] attr_accessor :rows # Additional statistics about a ResultSet or PartialResultSet. # Corresponds to the JSON property `stats` # @return [Google::Apis::SpannerV1::ResultSetStats] attr_accessor :stats def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @metadata = args[:metadata] if args.key?(:metadata) @rows = args[:rows] if args.key?(:rows) @stats = args[:stats] if args.key?(:stats) end end # Metadata about a ResultSet or PartialResultSet. class ResultSetMetadata include Google::Apis::Core::Hashable # `StructType` defines the fields of a STRUCT type. # Corresponds to the JSON property `rowType` # @return [Google::Apis::SpannerV1::StructType] attr_accessor :row_type # A transaction. # Corresponds to the JSON property `transaction` # @return [Google::Apis::SpannerV1::Transaction] attr_accessor :transaction def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @row_type = args[:row_type] if args.key?(:row_type) @transaction = args[:transaction] if args.key?(:transaction) end end # Additional statistics about a ResultSet or PartialResultSet. class ResultSetStats include Google::Apis::Core::Hashable # Contains an ordered list of nodes appearing in the query plan. # Corresponds to the JSON property `queryPlan` # @return [Google::Apis::SpannerV1::QueryPlan] attr_accessor :query_plan # Aggregated statistics from the execution of the query. Only present when the # query is profiled. For example, a query could return the statistics as follows: # ` "rows_returned": "3", "elapsed_time": "1.22 secs", "cpu_time": "1.19 secs" ` # Corresponds to the JSON property `queryStats` # @return [Hash] attr_accessor :query_stats # Standard DML returns an exact count of rows that were modified. # Corresponds to the JSON property `rowCountExact` # @return [Fixnum] attr_accessor :row_count_exact # Partitioned DML does not offer exactly-once semantics, so it returns a lower # bound of the rows modified. # Corresponds to the JSON property `rowCountLowerBound` # @return [Fixnum] attr_accessor :row_count_lower_bound def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @query_plan = args[:query_plan] if args.key?(:query_plan) @query_stats = args[:query_stats] if args.key?(:query_stats) @row_count_exact = args[:row_count_exact] if args.key?(:row_count_exact) @row_count_lower_bound = args[:row_count_lower_bound] if args.key?(:row_count_lower_bound) end end # The request for Rollback. class RollbackRequest include Google::Apis::Core::Hashable # Required. The transaction to roll back. # Corresponds to the JSON property `transactionId` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :transaction_id def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @transaction_id = args[:transaction_id] if args.key?(:transaction_id) end end # A session in the Cloud Spanner API. class Session include Google::Apis::Core::Hashable # Output only. The approximate timestamp when the session is last used. It is # typically earlier than the actual last use time. # Corresponds to the JSON property `approximateLastUseTime` # @return [String] attr_accessor :approximate_last_use_time # Output only. The timestamp when the session is created. # Corresponds to the JSON property `createTime` # @return [String] attr_accessor :create_time # The labels for the session. * Label keys must be between 1 and 63 characters # long and must conform to the following regular expression: `[a-z]([-a-z0-9]*[a- # z0-9])?`. * Label values must be between 0 and 63 characters long and must # conform to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. * No more # than 64 labels can be associated with a given session. See https://goo.gl/ # xmQnxf for more information on and examples of labels. # Corresponds to the JSON property `labels` # @return [Hash] attr_accessor :labels # Output only. The name of the session. This is always system-assigned. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @approximate_last_use_time = args[:approximate_last_use_time] if args.key?(:approximate_last_use_time) @create_time = args[:create_time] if args.key?(:create_time) @labels = args[:labels] if args.key?(:labels) @name = args[:name] if args.key?(:name) end end # Request message for `SetIamPolicy` method. class SetIamPolicyRequest include Google::Apis::Core::Hashable # An Identity and Access Management (IAM) policy, which specifies access # controls for Google Cloud resources. A `Policy` is a collection of `bindings`. # A `binding` binds one or more `members` to a single `role`. Members can be # user accounts, service accounts, Google groups, and domains (such as G Suite). # A `role` is a named list of permissions; each `role` can be an IAM predefined # role or a user-created custom role. For some types of Google Cloud resources, # a `binding` can also specify a `condition`, which is a logical expression that # allows access to a resource only if the expression evaluates to `true`. A # condition can add constraints based on attributes of the request, the resource, # or both. To learn which resources support conditions in their IAM policies, # see the [IAM documentation](https://cloud.google.com/iam/help/conditions/ # resource-policies). **JSON example:** ` "bindings": [ ` "role": "roles/ # resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", " # group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@ # appspot.gserviceaccount.com" ] `, ` "role": "roles/resourcemanager. # organizationViewer", "members": [ "user:eve@example.com" ], "condition": ` " # title": "expirable access", "description": "Does not grant access after Sep # 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", ` # ` ], "etag": "BwWWja0YfJA=", "version": 3 ` **YAML example:** bindings: - # members: - user:mike@example.com - group:admins@example.com - domain:google. # com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/ # resourcemanager.organizationAdmin - members: - user:eve@example.com role: # roles/resourcemanager.organizationViewer condition: title: expirable access # description: Does not grant access after Sep 2020 expression: request.time < # timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a # description of IAM and its features, see the [IAM documentation](https://cloud. # google.com/iam/docs/). # Corresponds to the JSON property `policy` # @return [Google::Apis::SpannerV1::Policy] attr_accessor :policy def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @policy = args[:policy] if args.key?(:policy) end end # Condensed representation of a node and its subtree. Only present for `SCALAR` # PlanNode(s). class ShortRepresentation include Google::Apis::Core::Hashable # A string representation of the expression subtree rooted at this node. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description # A mapping of (subquery variable name) -> (subquery node id) for cases where # the `description` string of this node references a `SCALAR` subquery contained # in the expression subtree rooted at this node. The referenced `SCALAR` # subquery may not necessarily be a direct child of this node. # Corresponds to the JSON property `subqueries` # @return [Hash] attr_accessor :subqueries def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @description = args[:description] if args.key?(:description) @subqueries = args[:subqueries] if args.key?(:subqueries) end end # A single DML statement. class Statement include Google::Apis::Core::Hashable # It is not always possible for Cloud Spanner to infer the right SQL type from a # JSON value. For example, values of type `BYTES` and values of type `STRING` # both appear in params as JSON strings. In these cases, `param_types` can be # used to specify the exact SQL type for some or all of the SQL statement # parameters. See the definition of Type for more information about SQL types. # Corresponds to the JSON property `paramTypes` # @return [Hash] attr_accessor :param_types # Parameter names and values that bind to placeholders in the DML string. A # parameter placeholder consists of the `@` character followed by the parameter # name (for example, `@firstName`). Parameter names can contain letters, numbers, # and underscores. Parameters can appear anywhere that a literal value is # expected. The same parameter name can be used more than once, for example: `" # WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL # statement with unbound parameters. # Corresponds to the JSON property `params` # @return [Hash] attr_accessor :params # Required. The DML string. # Corresponds to the JSON property `sql` # @return [String] attr_accessor :sql def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @param_types = args[:param_types] if args.key?(:param_types) @params = args[:params] if args.key?(:params) @sql = args[:sql] if args.key?(:sql) end end # The `Status` type defines a logical error model that is suitable for different # programming environments, including REST APIs and RPC APIs. It is used by [ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of # data: error code, error message, and error details. You can find out more # about this error model and how to work with it in the [API Design Guide](https: # //cloud.google.com/apis/design/errors). class Status include Google::Apis::Core::Hashable # The status code, which should be an enum value of google.rpc.Code. # Corresponds to the JSON property `code` # @return [Fixnum] attr_accessor :code # A list of messages that carry the error details. There is a common set of # message types for APIs to use. # Corresponds to the JSON property `details` # @return [Array>] attr_accessor :details # A developer-facing error message, which should be in English. Any user-facing # error message should be localized and sent in the google.rpc.Status.details # field, or localized by the client. # Corresponds to the JSON property `message` # @return [String] attr_accessor :message def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @code = args[:code] if args.key?(:code) @details = args[:details] if args.key?(:details) @message = args[:message] if args.key?(:message) end end # `StructType` defines the fields of a STRUCT type. class StructType include Google::Apis::Core::Hashable # The list of fields that make up this struct. Order is significant, because # values of this struct type are represented as lists, where the order of field # values matches the order of fields in the StructType. In turn, the order of # fields matches the order of columns in a read request, or the order of fields # in the `SELECT` clause of a query. # Corresponds to the JSON property `fields` # @return [Array] attr_accessor :fields def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @fields = args[:fields] if args.key?(:fields) end end # Request message for `TestIamPermissions` method. class TestIamPermissionsRequest include Google::Apis::Core::Hashable # REQUIRED: The set of permissions to check for 'resource'. Permissions with # wildcards (such as '*', 'spanner.*', 'spanner.instances.*') are not allowed. # Corresponds to the JSON property `permissions` # @return [Array] attr_accessor :permissions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @permissions = args[:permissions] if args.key?(:permissions) end end # Response message for `TestIamPermissions` method. class TestIamPermissionsResponse include Google::Apis::Core::Hashable # A subset of `TestPermissionsRequest.permissions` that the caller is allowed. # Corresponds to the JSON property `permissions` # @return [Array] attr_accessor :permissions def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @permissions = args[:permissions] if args.key?(:permissions) end end # A transaction. class Transaction include Google::Apis::Core::Hashable # `id` may be used to identify the transaction in subsequent Read, ExecuteSql, # Commit, or Rollback calls. Single-use read-only transactions do not have IDs, # because single-use transactions do not support multiple requests. # Corresponds to the JSON property `id` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :id # For snapshot read-only transactions, the read timestamp chosen for the # transaction. Not returned by default: see TransactionOptions.ReadOnly. # return_read_timestamp. A timestamp in RFC3339 UTC \"Zulu\" format, accurate to # nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`. # Corresponds to the JSON property `readTimestamp` # @return [String] attr_accessor :read_timestamp def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @id = args[:id] if args.key?(:id) @read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp) end end # # Transactions Each session can have at most one active transaction at a time ( # note that standalone reads and queries use a transaction internally and do # count towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. It # is not necessary to create a new session for each transaction. # Transaction # Modes Cloud Spanner supports three transaction modes: 1. Locking read-write. # This type of transaction is the only way to write data into Cloud Spanner. # These transactions rely on pessimistic locking and, if necessary, two-phase # commit. Locking read-write transactions may abort, requiring the application # to retry. 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow writes. Snapshot read- # only transactions can be configured to read at timestamps in the past. # Snapshot read-only transactions do not need to be committed. 3. Partitioned # DML. This type of transaction is used to execute a single Partitioned DML # statement. Partitioned DML partitions the key space and runs the DML statement # over each partition in parallel using separate, internal transactions that # commit independently. Partitioned DML transactions do not need to be committed. # For transactions that only read, snapshot read-only transactions provide # simpler semantics and are almost always faster. In particular, read-only # transactions do not take locks, so they do not conflict with read-write # transactions. As a consequence of not taking locks, they also do not abort, so # retry loops are not needed. Transactions may only read/write data in a single # database. They may, however, read/write data in different tables within that # database. ## Locking Read-Write Transactions Locking transactions may be used # to atomically read-modify-write data anywhere in a database. This type of # transaction is externally consistent. Clients should attempt to minimize the # amount of time a transaction is active. Faster transactions commit with higher # probability and cause less contention. Cloud Spanner attempts to keep read # locks active as long as the transaction continues to do reads, and the # transaction has not been terminated by Commit or Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a transaction's # locks and abort it. Conceptually, a read-write transaction consists of zero or # more reads or SQL statements followed by Commit. At any time before Commit, # the client can send a Rollback request to abort the transaction. ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired are # still valid at commit time, and it is able to acquire write locks for all # writes. Cloud Spanner can abort the transaction for any reason. If a commit # attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has # not modified any user data in Cloud Spanner. Unless the transaction commits, # Cloud Spanner makes no guarantees about how long the transaction's locks were # held for. It is an error to use Cloud Spanner locks for any sort of mutual # exclusion other than between Cloud Spanner transactions themselves. ### # Retrying Aborted Transactions When a transaction aborts, the application can # choose to retry the whole transaction again. To maximize the chances of # successfully committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock priority # increases with each consecutive abort, meaning that each attempt has a # slightly better chance of success than the previous. Under some circumstances ( # e.g., many transactions attempting to modify the same row(s)), a transaction # can abort many times in a short period before successfully committing. Thus, # it is not a good idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent retrying. ## # # Idle Transactions A transaction is considered idle if it has no outstanding # reads or SQL queries and has not started a read or SQL query within the last # 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don' # t hold on to locks indefinitely. In that case, the commit will fail with error # `ABORTED`. If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from # becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only # transactions provides a simpler method than locking read-write transactions # for doing several consistent reads. However, this type of transaction does not # support writes. Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. # Since they do not acquire locks, they do not block concurrent read-write # transactions. Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read timestamp is # garbage collected; however, the default garbage collection policy is generous # enough that most applications do not need to worry about this in practice. # Snapshot read-only transactions do not need to call Commit or Rollback (and in # fact are not permitted to do so). To execute a snapshot transaction, the # client specifies a timestamp bound, which tells Cloud Spanner how to choose a # read timestamp. The types of timestamp bound are: - Strong (the default). - # Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read # is geographically distributed, stale read-only transactions can execute more # quickly than strong or read-write transaction, because they are able to # execute far from the leader replica. Each type of timestamp bound is discussed # in detail below. ### Strong Strong reads are guaranteed to see the effects of # all transactions that have committed before the start of the read. Furthermore, # all rows yielded by a single read are consistent with each other -- if any # part of the read observes a transaction, all parts of the read see the # transaction. Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are concurrent writes. # If consistency across reads is required, the reads should be executed within a # transaction or at an exact read timestamp. See TransactionOptions.ReadOnly. # strong. ### Exact Staleness These timestamp bounds execute reads at a user- # specified timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe modifications done by # all transactions with a commit timestamp <= the read timestamp, and observe # none of the modifications done by transactions with a larger commit timestamp. # They will block until all conflicting transactions that may be assigned commit # timestamps <= the read timestamp have finished. The timestamp can either be # expressed as an absolute Cloud Spanner commit timestamp or a staleness # relative to the current time. These modes do not require a "negotiation phase" # to pick a timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, boundedly # stale reads usually return fresher results. See TransactionOptions.ReadOnly. # read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded # Staleness Bounded staleness modes allow Cloud Spanner to pick the read # timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses # the newest timestamp within the staleness bound that allows execution of the # reads at the closest available replica without blocking. All rows yielded are # consistent with each other -- if any part of the read observes a transaction, # all parts of the read see the transaction. Boundedly stale reads are not # repeatable: two stale reads, even if they use the same staleness bound, can # execute at different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase negotiates a # timestamp among all replicas needed to serve the read. In the second phase, # reads are executed at the negotiated timestamp. As a result of the two phase # execution, bounded staleness reads are usually a little slower than comparable # exact staleness reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. Because the # timestamp negotiation requires up-front knowledge of which rows will be read, # it can only be used with single-use read-only transactions. See # TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly. # min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud # Spanner continuously garbage collects deleted and overwritten data in the # background to reclaim storage space. This process is known as "version GC". By # default, version GC reclaims versions after they are one hour old. Because of # this, Cloud Spanner cannot perform reads at read timestamps more than one hour # in the past. This restriction also applies to in-progress reads and/or SQL # queries whose timestamp become too old while executing. Reads and SQL queries # with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ## # Partitioned DML Transactions Partitioned DML transactions are used to execute # DML statements with a different execution strategy that provides different, # and often better, scalability properties for large, table-wide operations than # DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP # workload, should prefer using ReadWrite transactions. Partitioned DML # partitions the keyspace and runs the DML statement on each partition in # separate, internal transactions. These transactions commit automatically when # complete, and run independently from one another. To reduce lock contention, # this execution strategy only acquires read locks on rows that match the WHERE # clause of the statement. Additionally, the smaller per-partition transactions # hold locks for less time. That said, Partitioned DML is not a drop-in # replacement for standard DML used in ReadWrite transactions. - The DML # statement must be fully-partitionable. Specifically, the statement must be # expressible as the union of many statements which each access only a single # row of the table. - The statement is not applied atomically to all rows of the # table. Rather, the statement is applied atomically to partitions of the table, # in independent transactions. Secondary index rows are updated atomically with # the base table rows. - Partitioned DML does not guarantee exactly-once # execution semantics against a partition. The statement will be applied at # least once to each partition. It is strongly recommended that the DML # statement should be idempotent to avoid unexpected results. For instance, it # is potentially dangerous to run a statement such as `UPDATE table SET column = # column + 1` as it could be run multiple times against some rows. - The # partitions are committed automatically - there is no support for Commit or # Rollback. If the call returns an error, or if the client issuing the # ExecuteSql call dies, it is possible that some rows had the statement executed # on them successfully. It is also possible that statement was never executed # against other rows. - Partitioned DML transactions may only contain the # execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - # If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the operation is # stopped at that point and an error is returned. It is possible that at this # point, some partitions have been committed (or even committed multiple times), # and other partitions have not been run at all. Given the above, Partitioned # DML is good fit for large, database-wide, operations that are idempotent, such # as deleting old rows from a very large table. class TransactionOptions include Google::Apis::Core::Hashable # Message type to initiate a Partitioned DML transaction. # Corresponds to the JSON property `partitionedDml` # @return [Google::Apis::SpannerV1::PartitionedDml] attr_accessor :partitioned_dml # Message type to initiate a read-only transaction. # Corresponds to the JSON property `readOnly` # @return [Google::Apis::SpannerV1::ReadOnly] attr_accessor :read_only # Message type to initiate a read-write transaction. Currently this transaction # type has no options. # Corresponds to the JSON property `readWrite` # @return [Google::Apis::SpannerV1::ReadWrite] attr_accessor :read_write def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @partitioned_dml = args[:partitioned_dml] if args.key?(:partitioned_dml) @read_only = args[:read_only] if args.key?(:read_only) @read_write = args[:read_write] if args.key?(:read_write) end end # This message is used to select the transaction in which a Read or ExecuteSql # call runs. See TransactionOptions for more information about transactions. class TransactionSelector include Google::Apis::Core::Hashable # # Transactions Each session can have at most one active transaction at a time ( # note that standalone reads and queries use a transaction internally and do # count towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. It # is not necessary to create a new session for each transaction. # Transaction # Modes Cloud Spanner supports three transaction modes: 1. Locking read-write. # This type of transaction is the only way to write data into Cloud Spanner. # These transactions rely on pessimistic locking and, if necessary, two-phase # commit. Locking read-write transactions may abort, requiring the application # to retry. 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow writes. Snapshot read- # only transactions can be configured to read at timestamps in the past. # Snapshot read-only transactions do not need to be committed. 3. Partitioned # DML. This type of transaction is used to execute a single Partitioned DML # statement. Partitioned DML partitions the key space and runs the DML statement # over each partition in parallel using separate, internal transactions that # commit independently. Partitioned DML transactions do not need to be committed. # For transactions that only read, snapshot read-only transactions provide # simpler semantics and are almost always faster. In particular, read-only # transactions do not take locks, so they do not conflict with read-write # transactions. As a consequence of not taking locks, they also do not abort, so # retry loops are not needed. Transactions may only read/write data in a single # database. They may, however, read/write data in different tables within that # database. ## Locking Read-Write Transactions Locking transactions may be used # to atomically read-modify-write data anywhere in a database. This type of # transaction is externally consistent. Clients should attempt to minimize the # amount of time a transaction is active. Faster transactions commit with higher # probability and cause less contention. Cloud Spanner attempts to keep read # locks active as long as the transaction continues to do reads, and the # transaction has not been terminated by Commit or Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a transaction's # locks and abort it. Conceptually, a read-write transaction consists of zero or # more reads or SQL statements followed by Commit. At any time before Commit, # the client can send a Rollback request to abort the transaction. ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired are # still valid at commit time, and it is able to acquire write locks for all # writes. Cloud Spanner can abort the transaction for any reason. If a commit # attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has # not modified any user data in Cloud Spanner. Unless the transaction commits, # Cloud Spanner makes no guarantees about how long the transaction's locks were # held for. It is an error to use Cloud Spanner locks for any sort of mutual # exclusion other than between Cloud Spanner transactions themselves. ### # Retrying Aborted Transactions When a transaction aborts, the application can # choose to retry the whole transaction again. To maximize the chances of # successfully committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock priority # increases with each consecutive abort, meaning that each attempt has a # slightly better chance of success than the previous. Under some circumstances ( # e.g., many transactions attempting to modify the same row(s)), a transaction # can abort many times in a short period before successfully committing. Thus, # it is not a good idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent retrying. ## # # Idle Transactions A transaction is considered idle if it has no outstanding # reads or SQL queries and has not started a read or SQL query within the last # 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don' # t hold on to locks indefinitely. In that case, the commit will fail with error # `ABORTED`. If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from # becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only # transactions provides a simpler method than locking read-write transactions # for doing several consistent reads. However, this type of transaction does not # support writes. Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. # Since they do not acquire locks, they do not block concurrent read-write # transactions. Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read timestamp is # garbage collected; however, the default garbage collection policy is generous # enough that most applications do not need to worry about this in practice. # Snapshot read-only transactions do not need to call Commit or Rollback (and in # fact are not permitted to do so). To execute a snapshot transaction, the # client specifies a timestamp bound, which tells Cloud Spanner how to choose a # read timestamp. The types of timestamp bound are: - Strong (the default). - # Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read # is geographically distributed, stale read-only transactions can execute more # quickly than strong or read-write transaction, because they are able to # execute far from the leader replica. Each type of timestamp bound is discussed # in detail below. ### Strong Strong reads are guaranteed to see the effects of # all transactions that have committed before the start of the read. Furthermore, # all rows yielded by a single read are consistent with each other -- if any # part of the read observes a transaction, all parts of the read see the # transaction. Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are concurrent writes. # If consistency across reads is required, the reads should be executed within a # transaction or at an exact read timestamp. See TransactionOptions.ReadOnly. # strong. ### Exact Staleness These timestamp bounds execute reads at a user- # specified timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe modifications done by # all transactions with a commit timestamp <= the read timestamp, and observe # none of the modifications done by transactions with a larger commit timestamp. # They will block until all conflicting transactions that may be assigned commit # timestamps <= the read timestamp have finished. The timestamp can either be # expressed as an absolute Cloud Spanner commit timestamp or a staleness # relative to the current time. These modes do not require a "negotiation phase" # to pick a timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, boundedly # stale reads usually return fresher results. See TransactionOptions.ReadOnly. # read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded # Staleness Bounded staleness modes allow Cloud Spanner to pick the read # timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses # the newest timestamp within the staleness bound that allows execution of the # reads at the closest available replica without blocking. All rows yielded are # consistent with each other -- if any part of the read observes a transaction, # all parts of the read see the transaction. Boundedly stale reads are not # repeatable: two stale reads, even if they use the same staleness bound, can # execute at different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase negotiates a # timestamp among all replicas needed to serve the read. In the second phase, # reads are executed at the negotiated timestamp. As a result of the two phase # execution, bounded staleness reads are usually a little slower than comparable # exact staleness reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. Because the # timestamp negotiation requires up-front knowledge of which rows will be read, # it can only be used with single-use read-only transactions. See # TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly. # min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud # Spanner continuously garbage collects deleted and overwritten data in the # background to reclaim storage space. This process is known as "version GC". By # default, version GC reclaims versions after they are one hour old. Because of # this, Cloud Spanner cannot perform reads at read timestamps more than one hour # in the past. This restriction also applies to in-progress reads and/or SQL # queries whose timestamp become too old while executing. Reads and SQL queries # with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ## # Partitioned DML Transactions Partitioned DML transactions are used to execute # DML statements with a different execution strategy that provides different, # and often better, scalability properties for large, table-wide operations than # DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP # workload, should prefer using ReadWrite transactions. Partitioned DML # partitions the keyspace and runs the DML statement on each partition in # separate, internal transactions. These transactions commit automatically when # complete, and run independently from one another. To reduce lock contention, # this execution strategy only acquires read locks on rows that match the WHERE # clause of the statement. Additionally, the smaller per-partition transactions # hold locks for less time. That said, Partitioned DML is not a drop-in # replacement for standard DML used in ReadWrite transactions. - The DML # statement must be fully-partitionable. Specifically, the statement must be # expressible as the union of many statements which each access only a single # row of the table. - The statement is not applied atomically to all rows of the # table. Rather, the statement is applied atomically to partitions of the table, # in independent transactions. Secondary index rows are updated atomically with # the base table rows. - Partitioned DML does not guarantee exactly-once # execution semantics against a partition. The statement will be applied at # least once to each partition. It is strongly recommended that the DML # statement should be idempotent to avoid unexpected results. For instance, it # is potentially dangerous to run a statement such as `UPDATE table SET column = # column + 1` as it could be run multiple times against some rows. - The # partitions are committed automatically - there is no support for Commit or # Rollback. If the call returns an error, or if the client issuing the # ExecuteSql call dies, it is possible that some rows had the statement executed # on them successfully. It is also possible that statement was never executed # against other rows. - Partitioned DML transactions may only contain the # execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - # If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the operation is # stopped at that point and an error is returned. It is possible that at this # point, some partitions have been committed (or even committed multiple times), # and other partitions have not been run at all. Given the above, Partitioned # DML is good fit for large, database-wide, operations that are idempotent, such # as deleting old rows from a very large table. # Corresponds to the JSON property `begin` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :begin # Execute the read or SQL query in a previously-started transaction. # Corresponds to the JSON property `id` # NOTE: Values are automatically base64 encoded/decoded in the client library. # @return [String] attr_accessor :id # # Transactions Each session can have at most one active transaction at a time ( # note that standalone reads and queries use a transaction internally and do # count towards the one transaction limit). After the active transaction is # completed, the session can immediately be re-used for the next transaction. It # is not necessary to create a new session for each transaction. # Transaction # Modes Cloud Spanner supports three transaction modes: 1. Locking read-write. # This type of transaction is the only way to write data into Cloud Spanner. # These transactions rely on pessimistic locking and, if necessary, two-phase # commit. Locking read-write transactions may abort, requiring the application # to retry. 2. Snapshot read-only. This transaction type provides guaranteed # consistency across several reads, but does not allow writes. Snapshot read- # only transactions can be configured to read at timestamps in the past. # Snapshot read-only transactions do not need to be committed. 3. Partitioned # DML. This type of transaction is used to execute a single Partitioned DML # statement. Partitioned DML partitions the key space and runs the DML statement # over each partition in parallel using separate, internal transactions that # commit independently. Partitioned DML transactions do not need to be committed. # For transactions that only read, snapshot read-only transactions provide # simpler semantics and are almost always faster. In particular, read-only # transactions do not take locks, so they do not conflict with read-write # transactions. As a consequence of not taking locks, they also do not abort, so # retry loops are not needed. Transactions may only read/write data in a single # database. They may, however, read/write data in different tables within that # database. ## Locking Read-Write Transactions Locking transactions may be used # to atomically read-modify-write data anywhere in a database. This type of # transaction is externally consistent. Clients should attempt to minimize the # amount of time a transaction is active. Faster transactions commit with higher # probability and cause less contention. Cloud Spanner attempts to keep read # locks active as long as the transaction continues to do reads, and the # transaction has not been terminated by Commit or Rollback. Long periods of # inactivity at the client may cause Cloud Spanner to release a transaction's # locks and abort it. Conceptually, a read-write transaction consists of zero or # more reads or SQL statements followed by Commit. At any time before Commit, # the client can send a Rollback request to abort the transaction. ### Semantics # Cloud Spanner can commit the transaction if all read locks it acquired are # still valid at commit time, and it is able to acquire write locks for all # writes. Cloud Spanner can abort the transaction for any reason. If a commit # attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has # not modified any user data in Cloud Spanner. Unless the transaction commits, # Cloud Spanner makes no guarantees about how long the transaction's locks were # held for. It is an error to use Cloud Spanner locks for any sort of mutual # exclusion other than between Cloud Spanner transactions themselves. ### # Retrying Aborted Transactions When a transaction aborts, the application can # choose to retry the whole transaction again. To maximize the chances of # successfully committing the retry, the client should execute the retry in the # same session as the original attempt. The original session's lock priority # increases with each consecutive abort, meaning that each attempt has a # slightly better chance of success than the previous. Under some circumstances ( # e.g., many transactions attempting to modify the same row(s)), a transaction # can abort many times in a short period before successfully committing. Thus, # it is not a good idea to cap the number of retries a transaction can attempt; # instead, it is better to limit the total amount of wall time spent retrying. ## # # Idle Transactions A transaction is considered idle if it has no outstanding # reads or SQL queries and has not started a read or SQL query within the last # 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don' # t hold on to locks indefinitely. In that case, the commit will fail with error # `ABORTED`. If this behavior is undesirable, periodically executing a simple # SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from # becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only # transactions provides a simpler method than locking read-write transactions # for doing several consistent reads. However, this type of transaction does not # support writes. Snapshot transactions do not take locks. Instead, they work by # choosing a Cloud Spanner timestamp, then executing all reads at that timestamp. # Since they do not acquire locks, they do not block concurrent read-write # transactions. Unlike locking read-write transactions, snapshot read-only # transactions never abort. They can fail if the chosen read timestamp is # garbage collected; however, the default garbage collection policy is generous # enough that most applications do not need to worry about this in practice. # Snapshot read-only transactions do not need to call Commit or Rollback (and in # fact are not permitted to do so). To execute a snapshot transaction, the # client specifies a timestamp bound, which tells Cloud Spanner how to choose a # read timestamp. The types of timestamp bound are: - Strong (the default). - # Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read # is geographically distributed, stale read-only transactions can execute more # quickly than strong or read-write transaction, because they are able to # execute far from the leader replica. Each type of timestamp bound is discussed # in detail below. ### Strong Strong reads are guaranteed to see the effects of # all transactions that have committed before the start of the read. Furthermore, # all rows yielded by a single read are consistent with each other -- if any # part of the read observes a transaction, all parts of the read see the # transaction. Strong reads are not repeatable: two consecutive strong read-only # transactions might return inconsistent results if there are concurrent writes. # If consistency across reads is required, the reads should be executed within a # transaction or at an exact read timestamp. See TransactionOptions.ReadOnly. # strong. ### Exact Staleness These timestamp bounds execute reads at a user- # specified timestamp. Reads at a timestamp are guaranteed to see a consistent # prefix of the global transaction history: they observe modifications done by # all transactions with a commit timestamp <= the read timestamp, and observe # none of the modifications done by transactions with a larger commit timestamp. # They will block until all conflicting transactions that may be assigned commit # timestamps <= the read timestamp have finished. The timestamp can either be # expressed as an absolute Cloud Spanner commit timestamp or a staleness # relative to the current time. These modes do not require a "negotiation phase" # to pick a timestamp. As a result, they execute slightly faster than the # equivalent boundedly stale concurrency modes. On the other hand, boundedly # stale reads usually return fresher results. See TransactionOptions.ReadOnly. # read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded # Staleness Bounded staleness modes allow Cloud Spanner to pick the read # timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses # the newest timestamp within the staleness bound that allows execution of the # reads at the closest available replica without blocking. All rows yielded are # consistent with each other -- if any part of the read observes a transaction, # all parts of the read see the transaction. Boundedly stale reads are not # repeatable: two stale reads, even if they use the same staleness bound, can # execute at different timestamps and thus return inconsistent results. # Boundedly stale reads execute in two phases: the first phase negotiates a # timestamp among all replicas needed to serve the read. In the second phase, # reads are executed at the negotiated timestamp. As a result of the two phase # execution, bounded staleness reads are usually a little slower than comparable # exact staleness reads. However, they are typically able to return fresher # results, and are more likely to execute at the closest replica. Because the # timestamp negotiation requires up-front knowledge of which rows will be read, # it can only be used with single-use read-only transactions. See # TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly. # min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud # Spanner continuously garbage collects deleted and overwritten data in the # background to reclaim storage space. This process is known as "version GC". By # default, version GC reclaims versions after they are one hour old. Because of # this, Cloud Spanner cannot perform reads at read timestamps more than one hour # in the past. This restriction also applies to in-progress reads and/or SQL # queries whose timestamp become too old while executing. Reads and SQL queries # with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ## # Partitioned DML Transactions Partitioned DML transactions are used to execute # DML statements with a different execution strategy that provides different, # and often better, scalability properties for large, table-wide operations than # DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP # workload, should prefer using ReadWrite transactions. Partitioned DML # partitions the keyspace and runs the DML statement on each partition in # separate, internal transactions. These transactions commit automatically when # complete, and run independently from one another. To reduce lock contention, # this execution strategy only acquires read locks on rows that match the WHERE # clause of the statement. Additionally, the smaller per-partition transactions # hold locks for less time. That said, Partitioned DML is not a drop-in # replacement for standard DML used in ReadWrite transactions. - The DML # statement must be fully-partitionable. Specifically, the statement must be # expressible as the union of many statements which each access only a single # row of the table. - The statement is not applied atomically to all rows of the # table. Rather, the statement is applied atomically to partitions of the table, # in independent transactions. Secondary index rows are updated atomically with # the base table rows. - Partitioned DML does not guarantee exactly-once # execution semantics against a partition. The statement will be applied at # least once to each partition. It is strongly recommended that the DML # statement should be idempotent to avoid unexpected results. For instance, it # is potentially dangerous to run a statement such as `UPDATE table SET column = # column + 1` as it could be run multiple times against some rows. - The # partitions are committed automatically - there is no support for Commit or # Rollback. If the call returns an error, or if the client issuing the # ExecuteSql call dies, it is possible that some rows had the statement executed # on them successfully. It is also possible that statement was never executed # against other rows. - Partitioned DML transactions may only contain the # execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. - # If any error is encountered during the execution of the partitioned DML # operation (for instance, a UNIQUE INDEX violation, division by zero, or a # value that cannot be stored due to schema constraints), then the operation is # stopped at that point and an error is returned. It is possible that at this # point, some partitions have been committed (or even committed multiple times), # and other partitions have not been run at all. Given the above, Partitioned # DML is good fit for large, database-wide, operations that are idempotent, such # as deleting old rows from a very large table. # Corresponds to the JSON property `singleUse` # @return [Google::Apis::SpannerV1::TransactionOptions] attr_accessor :single_use def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @begin = args[:begin] if args.key?(:begin) @id = args[:id] if args.key?(:id) @single_use = args[:single_use] if args.key?(:single_use) end end # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. class Type include Google::Apis::Core::Hashable # `Type` indicates the type of a Cloud Spanner value, as might be stored in a # table cell or returned from an SQL query. # Corresponds to the JSON property `arrayElementType` # @return [Google::Apis::SpannerV1::Type] attr_accessor :array_element_type # Required. The TypeCode for this type. # Corresponds to the JSON property `code` # @return [String] attr_accessor :code # `StructType` defines the fields of a STRUCT type. # Corresponds to the JSON property `structType` # @return [Google::Apis::SpannerV1::StructType] attr_accessor :struct_type def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @array_element_type = args[:array_element_type] if args.key?(:array_element_type) @code = args[:code] if args.key?(:code) @struct_type = args[:struct_type] if args.key?(:struct_type) end end # Metadata type for the operation returned by UpdateDatabaseDdl. class UpdateDatabaseDdlMetadata include Google::Apis::Core::Hashable # Reports the commit timestamps of all statements that have succeeded so far, # where `commit_timestamps[i]` is the commit timestamp for the statement ` # statements[i]`. # Corresponds to the JSON property `commitTimestamps` # @return [Array] attr_accessor :commit_timestamps # The database being modified. # Corresponds to the JSON property `database` # @return [String] attr_accessor :database # For an update this list contains all the statements. For an individual # statement, this list contains only that statement. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @commit_timestamps = args[:commit_timestamps] if args.key?(:commit_timestamps) @database = args[:database] if args.key?(:database) @statements = args[:statements] if args.key?(:statements) end end # Enqueues the given DDL statements to be applied, in order but not necessarily # all at once, to the database schema at some point (or points) in the future. # The server checks that the statements are executable (syntactically valid, # name tables that exist, etc.) before enqueueing them, but they may still fail # upon later execution (e.g., if a statement from another batch of statements is # applied first and it conflicts in some way, or if there is some data-related # problem like a `NULL` value in a column to which `NOT NULL` would be added). # If a statement fails, all subsequent statements in the batch are automatically # cancelled. Each batch of statements is assigned a name which can be used with # the Operations API to monitor progress. See the operation_id field for more # details. class UpdateDatabaseDdlRequest include Google::Apis::Core::Hashable # If empty, the new update request is assigned an automatically-generated # operation ID. Otherwise, `operation_id` is used to construct the name of the # resulting Operation. Specifying an explicit operation ID simplifies # determining whether the statements were executed in the event that the # UpdateDatabaseDdl call is replayed, or the return value is otherwise lost: the # database and `operation_id` fields can be combined to form the name of the # resulting longrunning.Operation: `/operations/`. `operation_id` should be # unique within the database, and must be a valid identifier: `a-z*`. Note that # automatically-generated operation IDs always begin with an underscore. If the # named operation already exists, UpdateDatabaseDdl returns `ALREADY_EXISTS`. # Corresponds to the JSON property `operationId` # @return [String] attr_accessor :operation_id # Required. DDL statements to be applied to the database. # Corresponds to the JSON property `statements` # @return [Array] attr_accessor :statements def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @operation_id = args[:operation_id] if args.key?(:operation_id) @statements = args[:statements] if args.key?(:statements) end end # Metadata type for the operation returned by UpdateInstance. class UpdateInstanceMetadata include Google::Apis::Core::Hashable # The time at which this operation was cancelled. If set, this operation is in # the process of undoing itself (which is guaranteed to succeed) and cannot be # cancelled again. # Corresponds to the JSON property `cancelTime` # @return [String] attr_accessor :cancel_time # The time at which this operation failed or was completed successfully. # Corresponds to the JSON property `endTime` # @return [String] attr_accessor :end_time # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance # The time at which UpdateInstance request was received. # Corresponds to the JSON property `startTime` # @return [String] attr_accessor :start_time def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @cancel_time = args[:cancel_time] if args.key?(:cancel_time) @end_time = args[:end_time] if args.key?(:end_time) @instance = args[:instance] if args.key?(:instance) @start_time = args[:start_time] if args.key?(:start_time) end end # The request for UpdateInstance. class UpdateInstanceRequest include Google::Apis::Core::Hashable # Required. A mask specifying which fields in Instance should be updated. The # field mask must always be specified; this prevents any future fields in # Instance from being erased accidentally by clients that do not know about them. # Corresponds to the JSON property `fieldMask` # @return [String] attr_accessor :field_mask # An isolated set of Cloud Spanner resources on which databases can be hosted. # Corresponds to the JSON property `instance` # @return [Google::Apis::SpannerV1::Instance] attr_accessor :instance def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @field_mask = args[:field_mask] if args.key?(:field_mask) @instance = args[:instance] if args.key?(:instance) end end # Arguments to insert, update, insert_or_update, and replace operations. class Write include Google::Apis::Core::Hashable # The names of the columns in table to be written. The list of columns must # contain enough columns to allow Cloud Spanner to derive values for all primary # key columns in the row(s) to be modified. # Corresponds to the JSON property `columns` # @return [Array] attr_accessor :columns # Required. The table whose rows will be written. # Corresponds to the JSON property `table` # @return [String] attr_accessor :table # The values to be written. `values` can contain more than one list of values. # If it does, then multiple rows are written, one for each entry in `values`. # Each list in `values` must have exactly as many entries as there are entries # in columns above. Sending multiple lists is equivalent to sending multiple ` # Mutation`s, each containing one `values` entry and repeating table and columns. # Individual values in each list are encoded as described here. # Corresponds to the JSON property `values` # @return [Array>] attr_accessor :values def initialize(**args) update!(**args) end # Update properties of this object def update!(**args) @columns = args[:columns] if args.key?(:columns) @table = args[:table] if args.key?(:table) @values = args[:values] if args.key?(:values) end end end end end