google-api-ruby-client/generated/google/apis/spanner_v1/classes.rb

3870 lines
202 KiB
Ruby

# Copyright 2015 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
require 'date'
require 'google/apis/core/base_service'
require 'google/apis/core/json_representation'
require 'google/apis/core/hashable'
require 'google/apis/errors'
module Google
module Apis
module SpannerV1
# A backup of a Cloud Spanner database.
class Backup
include Google::Apis::Core::Hashable
# Output only. The backup will contain an externally consistent copy of the
# database at the timestamp specified by `create_time`. `create_time` is
# approximately the time the CreateBackup request is received.
# Corresponds to the JSON property `createTime`
# @return [String]
attr_accessor :create_time
# Required for the CreateBackup operation. Name of the database from which this
# backup was created. This needs to be in the same instance as the backup.
# Values are of the form `projects//instances//databases/`.
# Corresponds to the JSON property `database`
# @return [String]
attr_accessor :database
# Required for the CreateBackup operation. The expiration time of the backup,
# with microseconds granularity that must be at least 6 hours and at most 366
# days from the time the CreateBackup request is processed. Once the `
# expire_time` has passed, the backup is eligible to be automatically deleted by
# Cloud Spanner to free the resources used by the backup.
# Corresponds to the JSON property `expireTime`
# @return [String]
attr_accessor :expire_time
# Output only for the CreateBackup operation. Required for the UpdateBackup
# operation. A globally unique identifier for the backup which cannot be changed.
# Values are of the form `projects//instances//backups/a-z*[a-z0-9]` The final
# segment of the name must be between 2 and 60 characters in length. The backup
# is stored in the location(s) specified in the instance configuration of the
# instance containing the backup, identified by the prefix of the backup name of
# the form `projects//instances/`.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# Output only. The names of the restored databases that reference the backup.
# The database names are of the form `projects//instances//databases/`.
# Referencing databases may exist in different instances. The existence of any
# referencing database prevents the backup from being deleted. When a restored
# database from the backup enters the `READY` state, the reference to the backup
# is removed.
# Corresponds to the JSON property `referencingDatabases`
# @return [Array<String>]
attr_accessor :referencing_databases
# Output only. Size of the backup in bytes.
# Corresponds to the JSON property `sizeBytes`
# @return [Fixnum]
attr_accessor :size_bytes
# Output only. The current state of the backup.
# Corresponds to the JSON property `state`
# @return [String]
attr_accessor :state
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@create_time = args[:create_time] if args.key?(:create_time)
@database = args[:database] if args.key?(:database)
@expire_time = args[:expire_time] if args.key?(:expire_time)
@name = args[:name] if args.key?(:name)
@referencing_databases = args[:referencing_databases] if args.key?(:referencing_databases)
@size_bytes = args[:size_bytes] if args.key?(:size_bytes)
@state = args[:state] if args.key?(:state)
end
end
# Information about a backup.
class BackupInfo
include Google::Apis::Core::Hashable
# Name of the backup.
# Corresponds to the JSON property `backup`
# @return [String]
attr_accessor :backup
# The backup contains an externally consistent copy of `source_database` at the
# timestamp specified by `create_time`.
# Corresponds to the JSON property `createTime`
# @return [String]
attr_accessor :create_time
# Name of the database the backup was created from.
# Corresponds to the JSON property `sourceDatabase`
# @return [String]
attr_accessor :source_database
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@backup = args[:backup] if args.key?(:backup)
@create_time = args[:create_time] if args.key?(:create_time)
@source_database = args[:source_database] if args.key?(:source_database)
end
end
# The request for BatchCreateSessions.
class BatchCreateSessionsRequest
include Google::Apis::Core::Hashable
# Required. The number of sessions to be created in this batch call. The API may
# return fewer than the requested number of sessions. If a specific number of
# sessions are desired, the client can make additional calls to
# BatchCreateSessions (adjusting session_count as necessary).
# Corresponds to the JSON property `sessionCount`
# @return [Fixnum]
attr_accessor :session_count
# A session in the Cloud Spanner API.
# Corresponds to the JSON property `sessionTemplate`
# @return [Google::Apis::SpannerV1::Session]
attr_accessor :session_template
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@session_count = args[:session_count] if args.key?(:session_count)
@session_template = args[:session_template] if args.key?(:session_template)
end
end
# The response for BatchCreateSessions.
class BatchCreateSessionsResponse
include Google::Apis::Core::Hashable
# The freshly created sessions.
# Corresponds to the JSON property `session`
# @return [Array<Google::Apis::SpannerV1::Session>]
attr_accessor :session
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@session = args[:session] if args.key?(:session)
end
end
# The request for BeginTransaction.
class BeginTransactionRequest
include Google::Apis::Core::Hashable
# # Transactions Each session can have at most one active transaction at a time (
# note that standalone reads and queries use a transaction internally and do
# count towards the one transaction limit). After the active transaction is
# completed, the session can immediately be re-used for the next transaction. It
# is not necessary to create a new session for each transaction. # Transaction
# Modes Cloud Spanner supports three transaction modes: 1. Locking read-write.
# This type of transaction is the only way to write data into Cloud Spanner.
# These transactions rely on pessimistic locking and, if necessary, two-phase
# commit. Locking read-write transactions may abort, requiring the application
# to retry. 2. Snapshot read-only. This transaction type provides guaranteed
# consistency across several reads, but does not allow writes. Snapshot read-
# only transactions can be configured to read at timestamps in the past.
# Snapshot read-only transactions do not need to be committed. 3. Partitioned
# DML. This type of transaction is used to execute a single Partitioned DML
# statement. Partitioned DML partitions the key space and runs the DML statement
# over each partition in parallel using separate, internal transactions that
# commit independently. Partitioned DML transactions do not need to be committed.
# For transactions that only read, snapshot read-only transactions provide
# simpler semantics and are almost always faster. In particular, read-only
# transactions do not take locks, so they do not conflict with read-write
# transactions. As a consequence of not taking locks, they also do not abort, so
# retry loops are not needed. Transactions may only read/write data in a single
# database. They may, however, read/write data in different tables within that
# database. ## Locking Read-Write Transactions Locking transactions may be used
# to atomically read-modify-write data anywhere in a database. This type of
# transaction is externally consistent. Clients should attempt to minimize the
# amount of time a transaction is active. Faster transactions commit with higher
# probability and cause less contention. Cloud Spanner attempts to keep read
# locks active as long as the transaction continues to do reads, and the
# transaction has not been terminated by Commit or Rollback. Long periods of
# inactivity at the client may cause Cloud Spanner to release a transaction's
# locks and abort it. Conceptually, a read-write transaction consists of zero or
# more reads or SQL statements followed by Commit. At any time before Commit,
# the client can send a Rollback request to abort the transaction. ### Semantics
# Cloud Spanner can commit the transaction if all read locks it acquired are
# still valid at commit time, and it is able to acquire write locks for all
# writes. Cloud Spanner can abort the transaction for any reason. If a commit
# attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has
# not modified any user data in Cloud Spanner. Unless the transaction commits,
# Cloud Spanner makes no guarantees about how long the transaction's locks were
# held for. It is an error to use Cloud Spanner locks for any sort of mutual
# exclusion other than between Cloud Spanner transactions themselves. ###
# Retrying Aborted Transactions When a transaction aborts, the application can
# choose to retry the whole transaction again. To maximize the chances of
# successfully committing the retry, the client should execute the retry in the
# same session as the original attempt. The original session's lock priority
# increases with each consecutive abort, meaning that each attempt has a
# slightly better chance of success than the previous. Under some circumstances (
# e.g., many transactions attempting to modify the same row(s)), a transaction
# can abort many times in a short period before successfully committing. Thus,
# it is not a good idea to cap the number of retries a transaction can attempt;
# instead, it is better to limit the total amount of wall time spent retrying. ##
# # Idle Transactions A transaction is considered idle if it has no outstanding
# reads or SQL queries and has not started a read or SQL query within the last
# 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don'
# t hold on to locks indefinitely. In that case, the commit will fail with error
# `ABORTED`. If this behavior is undesirable, periodically executing a simple
# SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from
# becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only
# transactions provides a simpler method than locking read-write transactions
# for doing several consistent reads. However, this type of transaction does not
# support writes. Snapshot transactions do not take locks. Instead, they work by
# choosing a Cloud Spanner timestamp, then executing all reads at that timestamp.
# Since they do not acquire locks, they do not block concurrent read-write
# transactions. Unlike locking read-write transactions, snapshot read-only
# transactions never abort. They can fail if the chosen read timestamp is
# garbage collected; however, the default garbage collection policy is generous
# enough that most applications do not need to worry about this in practice.
# Snapshot read-only transactions do not need to call Commit or Rollback (and in
# fact are not permitted to do so). To execute a snapshot transaction, the
# client specifies a timestamp bound, which tells Cloud Spanner how to choose a
# read timestamp. The types of timestamp bound are: - Strong (the default). -
# Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read
# is geographically distributed, stale read-only transactions can execute more
# quickly than strong or read-write transaction, because they are able to
# execute far from the leader replica. Each type of timestamp bound is discussed
# in detail below. ### Strong Strong reads are guaranteed to see the effects of
# all transactions that have committed before the start of the read. Furthermore,
# all rows yielded by a single read are consistent with each other -- if any
# part of the read observes a transaction, all parts of the read see the
# transaction. Strong reads are not repeatable: two consecutive strong read-only
# transactions might return inconsistent results if there are concurrent writes.
# If consistency across reads is required, the reads should be executed within a
# transaction or at an exact read timestamp. See TransactionOptions.ReadOnly.
# strong. ### Exact Staleness These timestamp bounds execute reads at a user-
# specified timestamp. Reads at a timestamp are guaranteed to see a consistent
# prefix of the global transaction history: they observe modifications done by
# all transactions with a commit timestamp <= the read timestamp, and observe
# none of the modifications done by transactions with a larger commit timestamp.
# They will block until all conflicting transactions that may be assigned commit
# timestamps <= the read timestamp have finished. The timestamp can either be
# expressed as an absolute Cloud Spanner commit timestamp or a staleness
# relative to the current time. These modes do not require a "negotiation phase"
# to pick a timestamp. As a result, they execute slightly faster than the
# equivalent boundedly stale concurrency modes. On the other hand, boundedly
# stale reads usually return fresher results. See TransactionOptions.ReadOnly.
# read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded
# Staleness Bounded staleness modes allow Cloud Spanner to pick the read
# timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses
# the newest timestamp within the staleness bound that allows execution of the
# reads at the closest available replica without blocking. All rows yielded are
# consistent with each other -- if any part of the read observes a transaction,
# all parts of the read see the transaction. Boundedly stale reads are not
# repeatable: two stale reads, even if they use the same staleness bound, can
# execute at different timestamps and thus return inconsistent results.
# Boundedly stale reads execute in two phases: the first phase negotiates a
# timestamp among all replicas needed to serve the read. In the second phase,
# reads are executed at the negotiated timestamp. As a result of the two phase
# execution, bounded staleness reads are usually a little slower than comparable
# exact staleness reads. However, they are typically able to return fresher
# results, and are more likely to execute at the closest replica. Because the
# timestamp negotiation requires up-front knowledge of which rows will be read,
# it can only be used with single-use read-only transactions. See
# TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.
# min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud
# Spanner continuously garbage collects deleted and overwritten data in the
# background to reclaim storage space. This process is known as "version GC". By
# default, version GC reclaims versions after they are one hour old. Because of
# this, Cloud Spanner cannot perform reads at read timestamps more than one hour
# in the past. This restriction also applies to in-progress reads and/or SQL
# queries whose timestamp become too old while executing. Reads and SQL queries
# with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ##
# Partitioned DML Transactions Partitioned DML transactions are used to execute
# DML statements with a different execution strategy that provides different,
# and often better, scalability properties for large, table-wide operations than
# DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP
# workload, should prefer using ReadWrite transactions. Partitioned DML
# partitions the keyspace and runs the DML statement on each partition in
# separate, internal transactions. These transactions commit automatically when
# complete, and run independently from one another. To reduce lock contention,
# this execution strategy only acquires read locks on rows that match the WHERE
# clause of the statement. Additionally, the smaller per-partition transactions
# hold locks for less time. That said, Partitioned DML is not a drop-in
# replacement for standard DML used in ReadWrite transactions. - The DML
# statement must be fully-partitionable. Specifically, the statement must be
# expressible as the union of many statements which each access only a single
# row of the table. - The statement is not applied atomically to all rows of the
# table. Rather, the statement is applied atomically to partitions of the table,
# in independent transactions. Secondary index rows are updated atomically with
# the base table rows. - Partitioned DML does not guarantee exactly-once
# execution semantics against a partition. The statement will be applied at
# least once to each partition. It is strongly recommended that the DML
# statement should be idempotent to avoid unexpected results. For instance, it
# is potentially dangerous to run a statement such as `UPDATE table SET column =
# column + 1` as it could be run multiple times against some rows. - The
# partitions are committed automatically - there is no support for Commit or
# Rollback. If the call returns an error, or if the client issuing the
# ExecuteSql call dies, it is possible that some rows had the statement executed
# on them successfully. It is also possible that statement was never executed
# against other rows. - Partitioned DML transactions may only contain the
# execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. -
# If any error is encountered during the execution of the partitioned DML
# operation (for instance, a UNIQUE INDEX violation, division by zero, or a
# value that cannot be stored due to schema constraints), then the operation is
# stopped at that point and an error is returned. It is possible that at this
# point, some partitions have been committed (or even committed multiple times),
# and other partitions have not been run at all. Given the above, Partitioned
# DML is good fit for large, database-wide, operations that are idempotent, such
# as deleting old rows from a very large table.
# Corresponds to the JSON property `options`
# @return [Google::Apis::SpannerV1::TransactionOptions]
attr_accessor :options
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@options = args[:options] if args.key?(:options)
end
end
# Associates `members` with a `role`.
class Binding
include Google::Apis::Core::Hashable
# Represents a textual expression in the Common Expression Language (CEL) syntax.
# CEL is a C-like expression language. The syntax and semantics of CEL are
# documented at https://github.com/google/cel-spec. Example (Comparison): title:
# "Summary size limit" description: "Determines if a summary is less than 100
# chars" expression: "document.summary.size() < 100" Example (Equality): title: "
# Requestor is owner" description: "Determines if requestor is the document
# owner" expression: "document.owner == request.auth.claims.email" Example (
# Logic): title: "Public documents" description: "Determine whether the document
# should be publicly visible" expression: "document.type != 'private' &&
# document.type != 'internal'" Example (Data Manipulation): title: "Notification
# string" description: "Create a notification string with a timestamp."
# expression: "'New message received at ' + string(document.create_time)" The
# exact variables and functions that may be referenced within an expression are
# determined by the service that evaluates it. See the service documentation for
# additional information.
# Corresponds to the JSON property `condition`
# @return [Google::Apis::SpannerV1::Expr]
attr_accessor :condition
# Specifies the identities requesting access for a Cloud Platform resource. `
# members` can have the following values: * `allUsers`: A special identifier
# that represents anyone who is on the internet; with or without a Google
# account. * `allAuthenticatedUsers`: A special identifier that represents
# anyone who is authenticated with a Google account or a service account. * `
# user:`emailid``: An email address that represents a specific Google account.
# For example, `alice@example.com` . * `serviceAccount:`emailid``: An email
# address that represents a service account. For example, `my-other-app@appspot.
# gserviceaccount.com`. * `group:`emailid``: An email address that represents a
# Google group. For example, `admins@example.com`. * `deleted:user:`emailid`?uid=
# `uniqueid``: An email address (plus unique identifier) representing a user
# that has been recently deleted. For example, `alice@example.com?uid=
# 123456789012345678901`. If the user is recovered, this value reverts to `user:`
# emailid`` and the recovered user retains the role in the binding. * `deleted:
# serviceAccount:`emailid`?uid=`uniqueid``: An email address (plus unique
# identifier) representing a service account that has been recently deleted. For
# example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`.
# If the service account is undeleted, this value reverts to `serviceAccount:`
# emailid`` and the undeleted service account retains the role in the binding. *
# `deleted:group:`emailid`?uid=`uniqueid``: An email address (plus unique
# identifier) representing a Google group that has been recently deleted. For
# example, `admins@example.com?uid=123456789012345678901`. If the group is
# recovered, this value reverts to `group:`emailid`` and the recovered group
# retains the role in the binding. * `domain:`domain``: The G Suite domain (
# primary) that represents all the users of that domain. For example, `google.
# com` or `example.com`.
# Corresponds to the JSON property `members`
# @return [Array<String>]
attr_accessor :members
# Role that is assigned to `members`. For example, `roles/viewer`, `roles/editor`
# , or `roles/owner`.
# Corresponds to the JSON property `role`
# @return [String]
attr_accessor :role
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@condition = args[:condition] if args.key?(:condition)
@members = args[:members] if args.key?(:members)
@role = args[:role] if args.key?(:role)
end
end
# Metadata associated with a parent-child relationship appearing in a PlanNode.
class ChildLink
include Google::Apis::Core::Hashable
# The node to which the link points.
# Corresponds to the JSON property `childIndex`
# @return [Fixnum]
attr_accessor :child_index
# The type of the link. For example, in Hash Joins this could be used to
# distinguish between the build child and the probe child, or in the case of the
# child being an output variable, to represent the tag associated with the
# output variable.
# Corresponds to the JSON property `type`
# @return [String]
attr_accessor :type
# Only present if the child node is SCALAR and corresponds to an output variable
# of the parent node. The field carries the name of the output variable. For
# example, a `TableScan` operator that reads rows from a table will have child
# links to the `SCALAR` nodes representing the output variables created for each
# column that is read by the operator. The corresponding `variable` fields will
# be set to the variable names assigned to the columns.
# Corresponds to the JSON property `variable`
# @return [String]
attr_accessor :variable
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@child_index = args[:child_index] if args.key?(:child_index)
@type = args[:type] if args.key?(:type)
@variable = args[:variable] if args.key?(:variable)
end
end
# The request for Commit.
class CommitRequest
include Google::Apis::Core::Hashable
# The mutations to be executed when this transaction commits. All mutations are
# applied atomically, in the order they appear in this list.
# Corresponds to the JSON property `mutations`
# @return [Array<Google::Apis::SpannerV1::Mutation>]
attr_accessor :mutations
# # Transactions Each session can have at most one active transaction at a time (
# note that standalone reads and queries use a transaction internally and do
# count towards the one transaction limit). After the active transaction is
# completed, the session can immediately be re-used for the next transaction. It
# is not necessary to create a new session for each transaction. # Transaction
# Modes Cloud Spanner supports three transaction modes: 1. Locking read-write.
# This type of transaction is the only way to write data into Cloud Spanner.
# These transactions rely on pessimistic locking and, if necessary, two-phase
# commit. Locking read-write transactions may abort, requiring the application
# to retry. 2. Snapshot read-only. This transaction type provides guaranteed
# consistency across several reads, but does not allow writes. Snapshot read-
# only transactions can be configured to read at timestamps in the past.
# Snapshot read-only transactions do not need to be committed. 3. Partitioned
# DML. This type of transaction is used to execute a single Partitioned DML
# statement. Partitioned DML partitions the key space and runs the DML statement
# over each partition in parallel using separate, internal transactions that
# commit independently. Partitioned DML transactions do not need to be committed.
# For transactions that only read, snapshot read-only transactions provide
# simpler semantics and are almost always faster. In particular, read-only
# transactions do not take locks, so they do not conflict with read-write
# transactions. As a consequence of not taking locks, they also do not abort, so
# retry loops are not needed. Transactions may only read/write data in a single
# database. They may, however, read/write data in different tables within that
# database. ## Locking Read-Write Transactions Locking transactions may be used
# to atomically read-modify-write data anywhere in a database. This type of
# transaction is externally consistent. Clients should attempt to minimize the
# amount of time a transaction is active. Faster transactions commit with higher
# probability and cause less contention. Cloud Spanner attempts to keep read
# locks active as long as the transaction continues to do reads, and the
# transaction has not been terminated by Commit or Rollback. Long periods of
# inactivity at the client may cause Cloud Spanner to release a transaction's
# locks and abort it. Conceptually, a read-write transaction consists of zero or
# more reads or SQL statements followed by Commit. At any time before Commit,
# the client can send a Rollback request to abort the transaction. ### Semantics
# Cloud Spanner can commit the transaction if all read locks it acquired are
# still valid at commit time, and it is able to acquire write locks for all
# writes. Cloud Spanner can abort the transaction for any reason. If a commit
# attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has
# not modified any user data in Cloud Spanner. Unless the transaction commits,
# Cloud Spanner makes no guarantees about how long the transaction's locks were
# held for. It is an error to use Cloud Spanner locks for any sort of mutual
# exclusion other than between Cloud Spanner transactions themselves. ###
# Retrying Aborted Transactions When a transaction aborts, the application can
# choose to retry the whole transaction again. To maximize the chances of
# successfully committing the retry, the client should execute the retry in the
# same session as the original attempt. The original session's lock priority
# increases with each consecutive abort, meaning that each attempt has a
# slightly better chance of success than the previous. Under some circumstances (
# e.g., many transactions attempting to modify the same row(s)), a transaction
# can abort many times in a short period before successfully committing. Thus,
# it is not a good idea to cap the number of retries a transaction can attempt;
# instead, it is better to limit the total amount of wall time spent retrying. ##
# # Idle Transactions A transaction is considered idle if it has no outstanding
# reads or SQL queries and has not started a read or SQL query within the last
# 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don'
# t hold on to locks indefinitely. In that case, the commit will fail with error
# `ABORTED`. If this behavior is undesirable, periodically executing a simple
# SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from
# becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only
# transactions provides a simpler method than locking read-write transactions
# for doing several consistent reads. However, this type of transaction does not
# support writes. Snapshot transactions do not take locks. Instead, they work by
# choosing a Cloud Spanner timestamp, then executing all reads at that timestamp.
# Since they do not acquire locks, they do not block concurrent read-write
# transactions. Unlike locking read-write transactions, snapshot read-only
# transactions never abort. They can fail if the chosen read timestamp is
# garbage collected; however, the default garbage collection policy is generous
# enough that most applications do not need to worry about this in practice.
# Snapshot read-only transactions do not need to call Commit or Rollback (and in
# fact are not permitted to do so). To execute a snapshot transaction, the
# client specifies a timestamp bound, which tells Cloud Spanner how to choose a
# read timestamp. The types of timestamp bound are: - Strong (the default). -
# Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read
# is geographically distributed, stale read-only transactions can execute more
# quickly than strong or read-write transaction, because they are able to
# execute far from the leader replica. Each type of timestamp bound is discussed
# in detail below. ### Strong Strong reads are guaranteed to see the effects of
# all transactions that have committed before the start of the read. Furthermore,
# all rows yielded by a single read are consistent with each other -- if any
# part of the read observes a transaction, all parts of the read see the
# transaction. Strong reads are not repeatable: two consecutive strong read-only
# transactions might return inconsistent results if there are concurrent writes.
# If consistency across reads is required, the reads should be executed within a
# transaction or at an exact read timestamp. See TransactionOptions.ReadOnly.
# strong. ### Exact Staleness These timestamp bounds execute reads at a user-
# specified timestamp. Reads at a timestamp are guaranteed to see a consistent
# prefix of the global transaction history: they observe modifications done by
# all transactions with a commit timestamp <= the read timestamp, and observe
# none of the modifications done by transactions with a larger commit timestamp.
# They will block until all conflicting transactions that may be assigned commit
# timestamps <= the read timestamp have finished. The timestamp can either be
# expressed as an absolute Cloud Spanner commit timestamp or a staleness
# relative to the current time. These modes do not require a "negotiation phase"
# to pick a timestamp. As a result, they execute slightly faster than the
# equivalent boundedly stale concurrency modes. On the other hand, boundedly
# stale reads usually return fresher results. See TransactionOptions.ReadOnly.
# read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded
# Staleness Bounded staleness modes allow Cloud Spanner to pick the read
# timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses
# the newest timestamp within the staleness bound that allows execution of the
# reads at the closest available replica without blocking. All rows yielded are
# consistent with each other -- if any part of the read observes a transaction,
# all parts of the read see the transaction. Boundedly stale reads are not
# repeatable: two stale reads, even if they use the same staleness bound, can
# execute at different timestamps and thus return inconsistent results.
# Boundedly stale reads execute in two phases: the first phase negotiates a
# timestamp among all replicas needed to serve the read. In the second phase,
# reads are executed at the negotiated timestamp. As a result of the two phase
# execution, bounded staleness reads are usually a little slower than comparable
# exact staleness reads. However, they are typically able to return fresher
# results, and are more likely to execute at the closest replica. Because the
# timestamp negotiation requires up-front knowledge of which rows will be read,
# it can only be used with single-use read-only transactions. See
# TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.
# min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud
# Spanner continuously garbage collects deleted and overwritten data in the
# background to reclaim storage space. This process is known as "version GC". By
# default, version GC reclaims versions after they are one hour old. Because of
# this, Cloud Spanner cannot perform reads at read timestamps more than one hour
# in the past. This restriction also applies to in-progress reads and/or SQL
# queries whose timestamp become too old while executing. Reads and SQL queries
# with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ##
# Partitioned DML Transactions Partitioned DML transactions are used to execute
# DML statements with a different execution strategy that provides different,
# and often better, scalability properties for large, table-wide operations than
# DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP
# workload, should prefer using ReadWrite transactions. Partitioned DML
# partitions the keyspace and runs the DML statement on each partition in
# separate, internal transactions. These transactions commit automatically when
# complete, and run independently from one another. To reduce lock contention,
# this execution strategy only acquires read locks on rows that match the WHERE
# clause of the statement. Additionally, the smaller per-partition transactions
# hold locks for less time. That said, Partitioned DML is not a drop-in
# replacement for standard DML used in ReadWrite transactions. - The DML
# statement must be fully-partitionable. Specifically, the statement must be
# expressible as the union of many statements which each access only a single
# row of the table. - The statement is not applied atomically to all rows of the
# table. Rather, the statement is applied atomically to partitions of the table,
# in independent transactions. Secondary index rows are updated atomically with
# the base table rows. - Partitioned DML does not guarantee exactly-once
# execution semantics against a partition. The statement will be applied at
# least once to each partition. It is strongly recommended that the DML
# statement should be idempotent to avoid unexpected results. For instance, it
# is potentially dangerous to run a statement such as `UPDATE table SET column =
# column + 1` as it could be run multiple times against some rows. - The
# partitions are committed automatically - there is no support for Commit or
# Rollback. If the call returns an error, or if the client issuing the
# ExecuteSql call dies, it is possible that some rows had the statement executed
# on them successfully. It is also possible that statement was never executed
# against other rows. - Partitioned DML transactions may only contain the
# execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. -
# If any error is encountered during the execution of the partitioned DML
# operation (for instance, a UNIQUE INDEX violation, division by zero, or a
# value that cannot be stored due to schema constraints), then the operation is
# stopped at that point and an error is returned. It is possible that at this
# point, some partitions have been committed (or even committed multiple times),
# and other partitions have not been run at all. Given the above, Partitioned
# DML is good fit for large, database-wide, operations that are idempotent, such
# as deleting old rows from a very large table.
# Corresponds to the JSON property `singleUseTransaction`
# @return [Google::Apis::SpannerV1::TransactionOptions]
attr_accessor :single_use_transaction
# Commit a previously-started transaction.
# Corresponds to the JSON property `transactionId`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :transaction_id
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@mutations = args[:mutations] if args.key?(:mutations)
@single_use_transaction = args[:single_use_transaction] if args.key?(:single_use_transaction)
@transaction_id = args[:transaction_id] if args.key?(:transaction_id)
end
end
# The response for Commit.
class CommitResponse
include Google::Apis::Core::Hashable
# The Cloud Spanner timestamp at which the transaction committed.
# Corresponds to the JSON property `commitTimestamp`
# @return [String]
attr_accessor :commit_timestamp
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@commit_timestamp = args[:commit_timestamp] if args.key?(:commit_timestamp)
end
end
# Metadata type for the operation returned by CreateBackup.
class CreateBackupMetadata
include Google::Apis::Core::Hashable
# The time at which cancellation of this operation was received. Operations.
# CancelOperation starts asynchronous cancellation on a long-running operation.
# The server makes a best effort to cancel the operation, but success is not
# guaranteed. Clients can use Operations.GetOperation or other methods to check
# whether the cancellation succeeded or whether the operation completed despite
# cancellation. On successful cancellation, the operation is not deleted;
# instead, it becomes an operation with an Operation.error value with a google.
# rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
# Corresponds to the JSON property `cancelTime`
# @return [String]
attr_accessor :cancel_time
# The name of the database the backup is created from.
# Corresponds to the JSON property `database`
# @return [String]
attr_accessor :database
# The name of the backup being created.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# Encapsulates progress related information for a Cloud Spanner long running
# operation.
# Corresponds to the JSON property `progress`
# @return [Google::Apis::SpannerV1::OperationProgress]
attr_accessor :progress
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@cancel_time = args[:cancel_time] if args.key?(:cancel_time)
@database = args[:database] if args.key?(:database)
@name = args[:name] if args.key?(:name)
@progress = args[:progress] if args.key?(:progress)
end
end
# Metadata type for the operation returned by CreateDatabase.
class CreateDatabaseMetadata
include Google::Apis::Core::Hashable
# The database being created.
# Corresponds to the JSON property `database`
# @return [String]
attr_accessor :database
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@database = args[:database] if args.key?(:database)
end
end
# The request for CreateDatabase.
class CreateDatabaseRequest
include Google::Apis::Core::Hashable
# Required. A `CREATE DATABASE` statement, which specifies the ID of the new
# database. The database ID must conform to the regular expression `a-z*[a-z0-9]`
# and be between 2 and 30 characters in length. If the database ID is a
# reserved word or if it contains a hyphen, the database ID must be enclosed in
# backticks (`` ` ``).
# Corresponds to the JSON property `createStatement`
# @return [String]
attr_accessor :create_statement
# Optional. A list of DDL statements to run inside the newly created database.
# Statements can create tables, indexes, etc. These statements execute
# atomically with the creation of the database: if there is an error in any
# statement, the database is not created.
# Corresponds to the JSON property `extraStatements`
# @return [Array<String>]
attr_accessor :extra_statements
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@create_statement = args[:create_statement] if args.key?(:create_statement)
@extra_statements = args[:extra_statements] if args.key?(:extra_statements)
end
end
# Metadata type for the operation returned by CreateInstance.
class CreateInstanceMetadata
include Google::Apis::Core::Hashable
# The time at which this operation was cancelled. If set, this operation is in
# the process of undoing itself (which is guaranteed to succeed) and cannot be
# cancelled again.
# Corresponds to the JSON property `cancelTime`
# @return [String]
attr_accessor :cancel_time
# The time at which this operation failed or was completed successfully.
# Corresponds to the JSON property `endTime`
# @return [String]
attr_accessor :end_time
# An isolated set of Cloud Spanner resources on which databases can be hosted.
# Corresponds to the JSON property `instance`
# @return [Google::Apis::SpannerV1::Instance]
attr_accessor :instance
# The time at which the CreateInstance request was received.
# Corresponds to the JSON property `startTime`
# @return [String]
attr_accessor :start_time
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@cancel_time = args[:cancel_time] if args.key?(:cancel_time)
@end_time = args[:end_time] if args.key?(:end_time)
@instance = args[:instance] if args.key?(:instance)
@start_time = args[:start_time] if args.key?(:start_time)
end
end
# The request for CreateInstance.
class CreateInstanceRequest
include Google::Apis::Core::Hashable
# An isolated set of Cloud Spanner resources on which databases can be hosted.
# Corresponds to the JSON property `instance`
# @return [Google::Apis::SpannerV1::Instance]
attr_accessor :instance
# Required. The ID of the instance to create. Valid identifiers are of the form `
# a-z*[a-z0-9]` and must be between 2 and 64 characters in length.
# Corresponds to the JSON property `instanceId`
# @return [String]
attr_accessor :instance_id
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@instance = args[:instance] if args.key?(:instance)
@instance_id = args[:instance_id] if args.key?(:instance_id)
end
end
# The request for CreateSession.
class CreateSessionRequest
include Google::Apis::Core::Hashable
# A session in the Cloud Spanner API.
# Corresponds to the JSON property `session`
# @return [Google::Apis::SpannerV1::Session]
attr_accessor :session
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@session = args[:session] if args.key?(:session)
end
end
# A Cloud Spanner database.
class Database
include Google::Apis::Core::Hashable
# Output only. If exists, the time at which the database creation started.
# Corresponds to the JSON property `createTime`
# @return [String]
attr_accessor :create_time
# Required. The name of the database. Values are of the form `projects//
# instances//databases/`, where `` is as specified in the `CREATE DATABASE`
# statement. This name can be passed to other API methods to identify the
# database.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# Information about the database restore.
# Corresponds to the JSON property `restoreInfo`
# @return [Google::Apis::SpannerV1::RestoreInfo]
attr_accessor :restore_info
# Output only. The current database state.
# Corresponds to the JSON property `state`
# @return [String]
attr_accessor :state
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@create_time = args[:create_time] if args.key?(:create_time)
@name = args[:name] if args.key?(:name)
@restore_info = args[:restore_info] if args.key?(:restore_info)
@state = args[:state] if args.key?(:state)
end
end
# Arguments to delete operations.
class Delete
include Google::Apis::Core::Hashable
# `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the
# keys are expected to be in the same table or index. The keys need not be
# sorted in any particular way. If the same key is specified multiple times in
# the set (for example if two ranges, two keys, or a key and a range overlap),
# Cloud Spanner behaves as if the key were only specified once.
# Corresponds to the JSON property `keySet`
# @return [Google::Apis::SpannerV1::KeySet]
attr_accessor :key_set
# Required. The table whose rows will be deleted.
# Corresponds to the JSON property `table`
# @return [String]
attr_accessor :table
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@key_set = args[:key_set] if args.key?(:key_set)
@table = args[:table] if args.key?(:table)
end
end
# A generic empty message that you can re-use to avoid defining duplicated empty
# messages in your APIs. A typical example is to use it as the request or the
# response type of an API method. For instance: service Foo ` rpc Bar(google.
# protobuf.Empty) returns (google.protobuf.Empty); ` The JSON representation for
# `Empty` is empty JSON object ````.
class Empty
include Google::Apis::Core::Hashable
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
end
end
# The request for ExecuteBatchDml.
class ExecuteBatchDmlRequest
include Google::Apis::Core::Hashable
# Required. A per-transaction sequence number used to identify this request.
# This field makes each request idempotent such that if the request is received
# multiple times, at most one will succeed. The sequence number must be
# monotonically increasing within the transaction. If a request arrives for the
# first time with an out-of-order sequence number, the transaction may be
# aborted. Replays of previously handled requests will yield the same response
# as the first execution.
# Corresponds to the JSON property `seqno`
# @return [Fixnum]
attr_accessor :seqno
# Required. The list of statements to execute in this batch. Statements are
# executed serially, such that the effects of statement `i` are visible to
# statement `i+1`. Each statement must be a DML statement. Execution stops at
# the first failed statement; the remaining statements are not executed. Callers
# must provide at least one statement.
# Corresponds to the JSON property `statements`
# @return [Array<Google::Apis::SpannerV1::Statement>]
attr_accessor :statements
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::TransactionSelector]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@seqno = args[:seqno] if args.key?(:seqno)
@statements = args[:statements] if args.key?(:statements)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# The response for ExecuteBatchDml. Contains a list of ResultSet messages, one
# for each DML statement that has successfully executed, in the same order as
# the statements in the request. If a statement fails, the status in the
# response body identifies the cause of the failure. To check for DML statements
# that failed, use the following approach: 1. Check the status in the response
# message. The google.rpc.Code enum value `OK` indicates that all statements
# were executed successfully. 2. If the status was not `OK`, check the number of
# result sets in the response. If the response contains `N` ResultSet messages,
# then statement `N+1` in the request failed. Example 1: * Request: 5 DML
# statements, all executed successfully. * Response: 5 ResultSet messages, with
# the status `OK`. Example 2: * Request: 5 DML statements. The third statement
# has a syntax error. * Response: 2 ResultSet messages, and a syntax error (`
# INVALID_ARGUMENT`) status. The number of ResultSet messages indicates that the
# third statement failed, and the fourth and fifth statements were not executed.
class ExecuteBatchDmlResponse
include Google::Apis::Core::Hashable
# One ResultSet for each statement in the request that ran successfully, in the
# same order as the statements in the request. Each ResultSet does not contain
# any rows. The ResultSetStats in each ResultSet contain the number of rows
# modified by the statement. Only the first ResultSet in the response contains
# valid ResultSetMetadata.
# Corresponds to the JSON property `resultSets`
# @return [Array<Google::Apis::SpannerV1::ResultSet>]
attr_accessor :result_sets
# The `Status` type defines a logical error model that is suitable for different
# programming environments, including REST APIs and RPC APIs. It is used by [
# gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
# data: error code, error message, and error details. You can find out more
# about this error model and how to work with it in the [API Design Guide](https:
# //cloud.google.com/apis/design/errors).
# Corresponds to the JSON property `status`
# @return [Google::Apis::SpannerV1::Status]
attr_accessor :status
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@result_sets = args[:result_sets] if args.key?(:result_sets)
@status = args[:status] if args.key?(:status)
end
end
# The request for ExecuteSql and ExecuteStreamingSql.
class ExecuteSqlRequest
include Google::Apis::Core::Hashable
# It is not always possible for Cloud Spanner to infer the right SQL type from a
# JSON value. For example, values of type `BYTES` and values of type `STRING`
# both appear in params as JSON strings. In these cases, `param_types` can be
# used to specify the exact SQL type for some or all of the SQL statement
# parameters. See the definition of Type for more information about SQL types.
# Corresponds to the JSON property `paramTypes`
# @return [Hash<String,Google::Apis::SpannerV1::Type>]
attr_accessor :param_types
# Parameter names and values that bind to placeholders in the SQL string. A
# parameter placeholder consists of the `@` character followed by the parameter
# name (for example, `@firstName`). Parameter names must conform to the naming
# requirements of identifiers as specified at https://cloud.google.com/spanner/
# docs/lexical#identifiers. Parameters can appear anywhere that a literal value
# is expected. The same parameter name can be used more than once, for example: `
# "WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL
# statement with unbound parameters.
# Corresponds to the JSON property `params`
# @return [Hash<String,Object>]
attr_accessor :params
# If present, results will be restricted to the specified partition previously
# created using PartitionQuery(). There must be an exact match for the values of
# fields common to this message and the PartitionQueryRequest message used to
# create this partition_token.
# Corresponds to the JSON property `partitionToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :partition_token
# Used to control the amount of debugging information returned in ResultSetStats.
# If partition_token is set, query_mode can only be set to QueryMode.NORMAL.
# Corresponds to the JSON property `queryMode`
# @return [String]
attr_accessor :query_mode
# Query optimizer configuration.
# Corresponds to the JSON property `queryOptions`
# @return [Google::Apis::SpannerV1::QueryOptions]
attr_accessor :query_options
# If this request is resuming a previously interrupted SQL statement execution, `
# resume_token` should be copied from the last PartialResultSet yielded before
# the interruption. Doing this enables the new SQL statement execution to resume
# where the last one left off. The rest of the request parameters must exactly
# match the request that yielded this token.
# Corresponds to the JSON property `resumeToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :resume_token
# A per-transaction sequence number used to identify this request. This field
# makes each request idempotent such that if the request is received multiple
# times, at most one will succeed. The sequence number must be monotonically
# increasing within the transaction. If a request arrives for the first time
# with an out-of-order sequence number, the transaction may be aborted. Replays
# of previously handled requests will yield the same response as the first
# execution. Required for DML statements. Ignored for queries.
# Corresponds to the JSON property `seqno`
# @return [Fixnum]
attr_accessor :seqno
# Required. The SQL string.
# Corresponds to the JSON property `sql`
# @return [String]
attr_accessor :sql
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::TransactionSelector]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@param_types = args[:param_types] if args.key?(:param_types)
@params = args[:params] if args.key?(:params)
@partition_token = args[:partition_token] if args.key?(:partition_token)
@query_mode = args[:query_mode] if args.key?(:query_mode)
@query_options = args[:query_options] if args.key?(:query_options)
@resume_token = args[:resume_token] if args.key?(:resume_token)
@seqno = args[:seqno] if args.key?(:seqno)
@sql = args[:sql] if args.key?(:sql)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# Represents a textual expression in the Common Expression Language (CEL) syntax.
# CEL is a C-like expression language. The syntax and semantics of CEL are
# documented at https://github.com/google/cel-spec. Example (Comparison): title:
# "Summary size limit" description: "Determines if a summary is less than 100
# chars" expression: "document.summary.size() < 100" Example (Equality): title: "
# Requestor is owner" description: "Determines if requestor is the document
# owner" expression: "document.owner == request.auth.claims.email" Example (
# Logic): title: "Public documents" description: "Determine whether the document
# should be publicly visible" expression: "document.type != 'private' &&
# document.type != 'internal'" Example (Data Manipulation): title: "Notification
# string" description: "Create a notification string with a timestamp."
# expression: "'New message received at ' + string(document.create_time)" The
# exact variables and functions that may be referenced within an expression are
# determined by the service that evaluates it. See the service documentation for
# additional information.
class Expr
include Google::Apis::Core::Hashable
# Optional. Description of the expression. This is a longer text which describes
# the expression, e.g. when hovered over it in a UI.
# Corresponds to the JSON property `description`
# @return [String]
attr_accessor :description
# Textual representation of an expression in Common Expression Language syntax.
# Corresponds to the JSON property `expression`
# @return [String]
attr_accessor :expression
# Optional. String indicating the location of the expression for error reporting,
# e.g. a file name and a position in the file.
# Corresponds to the JSON property `location`
# @return [String]
attr_accessor :location
# Optional. Title for the expression, i.e. a short string describing its purpose.
# This can be used e.g. in UIs which allow to enter the expression.
# Corresponds to the JSON property `title`
# @return [String]
attr_accessor :title
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@description = args[:description] if args.key?(:description)
@expression = args[:expression] if args.key?(:expression)
@location = args[:location] if args.key?(:location)
@title = args[:title] if args.key?(:title)
end
end
# Message representing a single field of a struct.
class Field
include Google::Apis::Core::Hashable
# The name of the field. For reads, this is the column name. For SQL queries, it
# is the column alias (e.g., `"Word"` in the query `"SELECT 'hello' AS Word"`),
# or the column name (e.g., `"ColName"` in the query `"SELECT ColName FROM Table"
# `). Some columns might have an empty name (e.g., `"SELECT UPPER(ColName)"`).
# Note that a query result can contain multiple fields with the same name.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# `Type` indicates the type of a Cloud Spanner value, as might be stored in a
# table cell or returned from an SQL query.
# Corresponds to the JSON property `type`
# @return [Google::Apis::SpannerV1::Type]
attr_accessor :type
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@name = args[:name] if args.key?(:name)
@type = args[:type] if args.key?(:type)
end
end
# The response for GetDatabaseDdl.
class GetDatabaseDdlResponse
include Google::Apis::Core::Hashable
# A list of formatted DDL statements defining the schema of the database
# specified in the request.
# Corresponds to the JSON property `statements`
# @return [Array<String>]
attr_accessor :statements
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@statements = args[:statements] if args.key?(:statements)
end
end
# Request message for `GetIamPolicy` method.
class GetIamPolicyRequest
include Google::Apis::Core::Hashable
# Encapsulates settings provided to GetIamPolicy.
# Corresponds to the JSON property `options`
# @return [Google::Apis::SpannerV1::GetPolicyOptions]
attr_accessor :options
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@options = args[:options] if args.key?(:options)
end
end
# Encapsulates settings provided to GetIamPolicy.
class GetPolicyOptions
include Google::Apis::Core::Hashable
# Optional. The policy format version to be returned. Valid values are 0, 1, and
# 3. Requests specifying an invalid value will be rejected. Requests for
# policies with any conditional bindings must specify version 3. Policies
# without any conditional bindings may specify any valid value or leave the
# field unset. To learn which resources support conditions in their IAM policies,
# see the [IAM documentation](https://cloud.google.com/iam/help/conditions/
# resource-policies).
# Corresponds to the JSON property `requestedPolicyVersion`
# @return [Fixnum]
attr_accessor :requested_policy_version
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@requested_policy_version = args[:requested_policy_version] if args.key?(:requested_policy_version)
end
end
# An isolated set of Cloud Spanner resources on which databases can be hosted.
class Instance
include Google::Apis::Core::Hashable
# Required. The name of the instance's configuration. Values are of the form `
# projects//instanceConfigs/`. See also InstanceConfig and ListInstanceConfigs.
# Corresponds to the JSON property `config`
# @return [String]
attr_accessor :config
# Required. The descriptive name for this instance as it appears in UIs. Must be
# unique per project and between 4 and 30 characters in length.
# Corresponds to the JSON property `displayName`
# @return [String]
attr_accessor :display_name
# Deprecated. This field is not populated.
# Corresponds to the JSON property `endpointUris`
# @return [Array<String>]
attr_accessor :endpoint_uris
# Cloud Labels are a flexible and lightweight mechanism for organizing cloud
# resources into groups that reflect a customer's organizational needs and
# deployment strategies. Cloud Labels can be used to filter collections of
# resources. They can be used to control how resource metrics are aggregated.
# And they can be used as arguments to policy management rules (e.g. route,
# firewall, load balancing, etc.). * Label keys must be between 1 and 63
# characters long and must conform to the following regular expression: `[a-z]([-
# a-z0-9]*[a-z0-9])?`. * Label values must be between 0 and 63 characters long
# and must conform to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. *
# No more than 64 labels can be associated with a given resource. See https://
# goo.gl/xmQnxf for more information on and examples of labels. If you plan to
# use labels in your own code, please note that additional characters may be
# allowed in the future. And so you are advised to use an internal label
# representation, such as JSON, which doesn't rely upon specific characters
# being disallowed. For example, representing labels as the string: name + "_" +
# value would prove problematic if we were to allow "_" in a future release.
# Corresponds to the JSON property `labels`
# @return [Hash<String,String>]
attr_accessor :labels
# Required. A unique identifier for the instance, which cannot be changed after
# the instance is created. Values are of the form `projects//instances/a-z*[a-z0-
# 9]`. The final segment of the name must be between 2 and 64 characters in
# length.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# The number of nodes allocated to this instance. This may be zero in API
# responses for instances that are not yet in state `READY`. See [the
# documentation](https://cloud.google.com/spanner/docs/instances#node_count) for
# more information about nodes.
# Corresponds to the JSON property `nodeCount`
# @return [Fixnum]
attr_accessor :node_count
# Output only. The current instance state. For CreateInstance, the state must be
# either omitted or set to `CREATING`. For UpdateInstance, the state must be
# either omitted or set to `READY`.
# Corresponds to the JSON property `state`
# @return [String]
attr_accessor :state
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@config = args[:config] if args.key?(:config)
@display_name = args[:display_name] if args.key?(:display_name)
@endpoint_uris = args[:endpoint_uris] if args.key?(:endpoint_uris)
@labels = args[:labels] if args.key?(:labels)
@name = args[:name] if args.key?(:name)
@node_count = args[:node_count] if args.key?(:node_count)
@state = args[:state] if args.key?(:state)
end
end
# A possible configuration for a Cloud Spanner instance. Configurations define
# the geographic placement of nodes and their replication.
class InstanceConfig
include Google::Apis::Core::Hashable
# The name of this instance configuration as it appears in UIs.
# Corresponds to the JSON property `displayName`
# @return [String]
attr_accessor :display_name
# A unique identifier for the instance configuration. Values are of the form `
# projects//instanceConfigs/a-z*`
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# The geographic placement of nodes in this instance configuration and their
# replication properties.
# Corresponds to the JSON property `replicas`
# @return [Array<Google::Apis::SpannerV1::ReplicaInfo>]
attr_accessor :replicas
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@display_name = args[:display_name] if args.key?(:display_name)
@name = args[:name] if args.key?(:name)
@replicas = args[:replicas] if args.key?(:replicas)
end
end
# KeyRange represents a range of rows in a table or index. A range has a start
# key and an end key. These keys can be open or closed, indicating if the range
# includes rows with that key. Keys are represented by lists, where the ith
# value in the list corresponds to the ith component of the table or index
# primary key. Individual values are encoded as described here. For example,
# consider the following table definition: CREATE TABLE UserEvents ( UserName
# STRING(MAX), EventDate STRING(10) ) PRIMARY KEY(UserName, EventDate); The
# following keys name rows in this table: "Bob", "2014-09-23" Since the `
# UserEvents` table's `PRIMARY KEY` clause names two columns, each `UserEvents`
# key has two elements; the first is the `UserName`, and the second is the `
# EventDate`. Key ranges with multiple components are interpreted
# lexicographically by component using the table or index key's declared sort
# order. For example, the following range returns all events for user `"Bob"`
# that occurred in the year 2015: "start_closed": ["Bob", "2015-01-01"] "
# end_closed": ["Bob", "2015-12-31"] Start and end keys can omit trailing key
# components. This affects the inclusion and exclusion of rows that exactly
# match the provided key components: if the key is closed, then rows that
# exactly match the provided components are included; if the key is open, then
# rows that exactly match are not included. For example, the following range
# includes all events for `"Bob"` that occurred during and after the year 2000: "
# start_closed": ["Bob", "2000-01-01"] "end_closed": ["Bob"] The next example
# retrieves all events for `"Bob"`: "start_closed": ["Bob"] "end_closed": ["Bob"]
# To retrieve events before the year 2000: "start_closed": ["Bob"] "end_open": [
# "Bob", "2000-01-01"] The following range includes all rows in the table: "
# start_closed": [] "end_closed": [] This range returns all users whose `
# UserName` begins with any character from A to C: "start_closed": ["A"] "
# end_open": ["D"] This range returns all users whose `UserName` begins with B: "
# start_closed": ["B"] "end_open": ["C"] Key ranges honor column sort order. For
# example, suppose a table is defined as follows: CREATE TABLE
# DescendingSortedTable ` Key INT64, ... ) PRIMARY KEY(Key DESC); The following
# range retrieves all rows with key values between 1 and 100 inclusive: "
# start_closed": ["100"] "end_closed": ["1"] Note that 100 is passed as the
# start, and 1 is passed as the end, because `Key` is a descending column in the
# schema.
class KeyRange
include Google::Apis::Core::Hashable
# If the end is closed, then the range includes all rows whose first `len(
# end_closed)` key columns exactly match `end_closed`.
# Corresponds to the JSON property `endClosed`
# @return [Array<Object>]
attr_accessor :end_closed
# If the end is open, then the range excludes rows whose first `len(end_open)`
# key columns exactly match `end_open`.
# Corresponds to the JSON property `endOpen`
# @return [Array<Object>]
attr_accessor :end_open
# If the start is closed, then the range includes all rows whose first `len(
# start_closed)` key columns exactly match `start_closed`.
# Corresponds to the JSON property `startClosed`
# @return [Array<Object>]
attr_accessor :start_closed
# If the start is open, then the range excludes rows whose first `len(start_open)
# ` key columns exactly match `start_open`.
# Corresponds to the JSON property `startOpen`
# @return [Array<Object>]
attr_accessor :start_open
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@end_closed = args[:end_closed] if args.key?(:end_closed)
@end_open = args[:end_open] if args.key?(:end_open)
@start_closed = args[:start_closed] if args.key?(:start_closed)
@start_open = args[:start_open] if args.key?(:start_open)
end
end
# `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the
# keys are expected to be in the same table or index. The keys need not be
# sorted in any particular way. If the same key is specified multiple times in
# the set (for example if two ranges, two keys, or a key and a range overlap),
# Cloud Spanner behaves as if the key were only specified once.
class KeySet
include Google::Apis::Core::Hashable
# For convenience `all` can be set to `true` to indicate that this `KeySet`
# matches all keys in the table or index. Note that any keys specified in `keys`
# or `ranges` are only yielded once.
# Corresponds to the JSON property `all`
# @return [Boolean]
attr_accessor :all
alias_method :all?, :all
# A list of specific keys. Entries in `keys` should have exactly as many
# elements as there are columns in the primary or index key with which this `
# KeySet` is used. Individual key values are encoded as described here.
# Corresponds to the JSON property `keys`
# @return [Array<Array<Object>>]
attr_accessor :keys
# A list of key ranges. See KeyRange for more information about key range
# specifications.
# Corresponds to the JSON property `ranges`
# @return [Array<Google::Apis::SpannerV1::KeyRange>]
attr_accessor :ranges
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@all = args[:all] if args.key?(:all)
@keys = args[:keys] if args.key?(:keys)
@ranges = args[:ranges] if args.key?(:ranges)
end
end
# The response for ListBackupOperations.
class ListBackupOperationsResponse
include Google::Apis::Core::Hashable
# `next_page_token` can be sent in a subsequent ListBackupOperations call to
# fetch more of the matching metadata.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
# The list of matching backup long-running operations. Each operation's name
# will be prefixed by the backup's name and the operation's metadata will be of
# type CreateBackupMetadata. Operations returned include those that are pending
# or have completed/failed/canceled within the last 7 days. Operations returned
# are ordered by `operation.metadata.value.progress.start_time` in descending
# order starting from the most recently started operation.
# Corresponds to the JSON property `operations`
# @return [Array<Google::Apis::SpannerV1::Operation>]
attr_accessor :operations
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
@operations = args[:operations] if args.key?(:operations)
end
end
# The response for ListBackups.
class ListBackupsResponse
include Google::Apis::Core::Hashable
# The list of matching backups. Backups returned are ordered by `create_time` in
# descending order, starting from the most recent `create_time`.
# Corresponds to the JSON property `backups`
# @return [Array<Google::Apis::SpannerV1::Backup>]
attr_accessor :backups
# `next_page_token` can be sent in a subsequent ListBackups call to fetch more
# of the matching backups.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@backups = args[:backups] if args.key?(:backups)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
end
end
# The response for ListDatabaseOperations.
class ListDatabaseOperationsResponse
include Google::Apis::Core::Hashable
# `next_page_token` can be sent in a subsequent ListDatabaseOperations call to
# fetch more of the matching metadata.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
# The list of matching database long-running operations. Each operation's name
# will be prefixed by the database's name. The operation's metadata field type `
# metadata.type_url` describes the type of the metadata.
# Corresponds to the JSON property `operations`
# @return [Array<Google::Apis::SpannerV1::Operation>]
attr_accessor :operations
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
@operations = args[:operations] if args.key?(:operations)
end
end
# The response for ListDatabases.
class ListDatabasesResponse
include Google::Apis::Core::Hashable
# Databases that matched the request.
# Corresponds to the JSON property `databases`
# @return [Array<Google::Apis::SpannerV1::Database>]
attr_accessor :databases
# `next_page_token` can be sent in a subsequent ListDatabases call to fetch more
# of the matching databases.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@databases = args[:databases] if args.key?(:databases)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
end
end
# The response for ListInstanceConfigs.
class ListInstanceConfigsResponse
include Google::Apis::Core::Hashable
# The list of requested instance configurations.
# Corresponds to the JSON property `instanceConfigs`
# @return [Array<Google::Apis::SpannerV1::InstanceConfig>]
attr_accessor :instance_configs
# `next_page_token` can be sent in a subsequent ListInstanceConfigs call to
# fetch more of the matching instance configurations.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@instance_configs = args[:instance_configs] if args.key?(:instance_configs)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
end
end
# The response for ListInstances.
class ListInstancesResponse
include Google::Apis::Core::Hashable
# The list of requested instances.
# Corresponds to the JSON property `instances`
# @return [Array<Google::Apis::SpannerV1::Instance>]
attr_accessor :instances
# `next_page_token` can be sent in a subsequent ListInstances call to fetch more
# of the matching instances.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@instances = args[:instances] if args.key?(:instances)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
end
end
# The response message for Operations.ListOperations.
class ListOperationsResponse
include Google::Apis::Core::Hashable
# The standard List next-page token.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
# A list of operations that matches the specified filter in the request.
# Corresponds to the JSON property `operations`
# @return [Array<Google::Apis::SpannerV1::Operation>]
attr_accessor :operations
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
@operations = args[:operations] if args.key?(:operations)
end
end
# The response for ListSessions.
class ListSessionsResponse
include Google::Apis::Core::Hashable
# `next_page_token` can be sent in a subsequent ListSessions call to fetch more
# of the matching sessions.
# Corresponds to the JSON property `nextPageToken`
# @return [String]
attr_accessor :next_page_token
# The list of requested sessions.
# Corresponds to the JSON property `sessions`
# @return [Array<Google::Apis::SpannerV1::Session>]
attr_accessor :sessions
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@next_page_token = args[:next_page_token] if args.key?(:next_page_token)
@sessions = args[:sessions] if args.key?(:sessions)
end
end
# A modification to one or more Cloud Spanner rows. Mutations can be applied to
# a Cloud Spanner database by sending them in a Commit call.
class Mutation
include Google::Apis::Core::Hashable
# Arguments to delete operations.
# Corresponds to the JSON property `delete`
# @return [Google::Apis::SpannerV1::Delete]
attr_accessor :delete
# Arguments to insert, update, insert_or_update, and replace operations.
# Corresponds to the JSON property `insert`
# @return [Google::Apis::SpannerV1::Write]
attr_accessor :insert
# Arguments to insert, update, insert_or_update, and replace operations.
# Corresponds to the JSON property `insertOrUpdate`
# @return [Google::Apis::SpannerV1::Write]
attr_accessor :insert_or_update
# Arguments to insert, update, insert_or_update, and replace operations.
# Corresponds to the JSON property `replace`
# @return [Google::Apis::SpannerV1::Write]
attr_accessor :replace
# Arguments to insert, update, insert_or_update, and replace operations.
# Corresponds to the JSON property `update`
# @return [Google::Apis::SpannerV1::Write]
attr_accessor :update
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@delete = args[:delete] if args.key?(:delete)
@insert = args[:insert] if args.key?(:insert)
@insert_or_update = args[:insert_or_update] if args.key?(:insert_or_update)
@replace = args[:replace] if args.key?(:replace)
@update = args[:update] if args.key?(:update)
end
end
# This resource represents a long-running operation that is the result of a
# network API call.
class Operation
include Google::Apis::Core::Hashable
# If the value is `false`, it means the operation is still in progress. If `true`
# , the operation is completed, and either `error` or `response` is available.
# Corresponds to the JSON property `done`
# @return [Boolean]
attr_accessor :done
alias_method :done?, :done
# The `Status` type defines a logical error model that is suitable for different
# programming environments, including REST APIs and RPC APIs. It is used by [
# gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
# data: error code, error message, and error details. You can find out more
# about this error model and how to work with it in the [API Design Guide](https:
# //cloud.google.com/apis/design/errors).
# Corresponds to the JSON property `error`
# @return [Google::Apis::SpannerV1::Status]
attr_accessor :error
# Service-specific metadata associated with the operation. It typically contains
# progress information and common metadata such as create time. Some services
# might not provide such metadata. Any method that returns a long-running
# operation should document the metadata type, if any.
# Corresponds to the JSON property `metadata`
# @return [Hash<String,Object>]
attr_accessor :metadata
# The server-assigned name, which is only unique within the same service that
# originally returns it. If you use the default HTTP mapping, the `name` should
# be a resource name ending with `operations/`unique_id``.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# The normal response of the operation in case of success. If the original
# method returns no data on success, such as `Delete`, the response is `google.
# protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`,
# the response should be the resource. For other methods, the response should
# have the type `XxxResponse`, where `Xxx` is the original method name. For
# example, if the original method name is `TakeSnapshot()`, the inferred
# response type is `TakeSnapshotResponse`.
# Corresponds to the JSON property `response`
# @return [Hash<String,Object>]
attr_accessor :response
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@done = args[:done] if args.key?(:done)
@error = args[:error] if args.key?(:error)
@metadata = args[:metadata] if args.key?(:metadata)
@name = args[:name] if args.key?(:name)
@response = args[:response] if args.key?(:response)
end
end
# Encapsulates progress related information for a Cloud Spanner long running
# operation.
class OperationProgress
include Google::Apis::Core::Hashable
# If set, the time at which this operation failed or was completed successfully.
# Corresponds to the JSON property `endTime`
# @return [String]
attr_accessor :end_time
# Percent completion of the operation. Values are between 0 and 100 inclusive.
# Corresponds to the JSON property `progressPercent`
# @return [Fixnum]
attr_accessor :progress_percent
# Time the request was received.
# Corresponds to the JSON property `startTime`
# @return [String]
attr_accessor :start_time
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@end_time = args[:end_time] if args.key?(:end_time)
@progress_percent = args[:progress_percent] if args.key?(:progress_percent)
@start_time = args[:start_time] if args.key?(:start_time)
end
end
# Metadata type for the long-running operation used to track the progress of
# optimizations performed on a newly restored database. This long-running
# operation is automatically created by the system after the successful
# completion of a database restore, and cannot be cancelled.
class OptimizeRestoredDatabaseMetadata
include Google::Apis::Core::Hashable
# Name of the restored database being optimized.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# Encapsulates progress related information for a Cloud Spanner long running
# operation.
# Corresponds to the JSON property `progress`
# @return [Google::Apis::SpannerV1::OperationProgress]
attr_accessor :progress
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@name = args[:name] if args.key?(:name)
@progress = args[:progress] if args.key?(:progress)
end
end
# Partial results from a streaming read or SQL query. Streaming reads and SQL
# queries better tolerate large result sets, large rows, and large values, but
# are a little trickier to consume.
class PartialResultSet
include Google::Apis::Core::Hashable
# If true, then the final value in values is chunked, and must be combined with
# more values from subsequent `PartialResultSet`s to obtain a complete field
# value.
# Corresponds to the JSON property `chunkedValue`
# @return [Boolean]
attr_accessor :chunked_value
alias_method :chunked_value?, :chunked_value
# Metadata about a ResultSet or PartialResultSet.
# Corresponds to the JSON property `metadata`
# @return [Google::Apis::SpannerV1::ResultSetMetadata]
attr_accessor :metadata
# Streaming calls might be interrupted for a variety of reasons, such as TCP
# connection loss. If this occurs, the stream of results can be resumed by re-
# sending the original request and including `resume_token`. Note that executing
# any other transaction in the same session invalidates the token.
# Corresponds to the JSON property `resumeToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :resume_token
# Additional statistics about a ResultSet or PartialResultSet.
# Corresponds to the JSON property `stats`
# @return [Google::Apis::SpannerV1::ResultSetStats]
attr_accessor :stats
# A streamed result set consists of a stream of values, which might be split
# into many `PartialResultSet` messages to accommodate large rows and/or large
# values. Every N complete values defines a row, where N is equal to the number
# of entries in metadata.row_type.fields. Most values are encoded based on type
# as described here. It is possible that the last value in values is "chunked",
# meaning that the rest of the value is sent in subsequent `PartialResultSet`(s).
# This is denoted by the chunked_value field. Two or more chunked values can be
# merged to form a complete value as follows: * `bool/number/null`: cannot be
# chunked * `string`: concatenate the strings * `list`: concatenate the lists.
# If the last element in a list is a `string`, `list`, or `object`, merge it
# with the first element in the next list by applying these rules recursively. *
# `object`: concatenate the (field name, field value) pairs. If a field name is
# duplicated, then apply these rules recursively to merge the field values. Some
# examples of merging: # Strings are concatenated. "foo", "bar" => "foobar" #
# Lists of non-strings are concatenated. [2, 3], [4] => [2, 3, 4] # Lists are
# concatenated, but the last and first elements are merged # because they are
# strings. ["a", "b"], ["c", "d"] => ["a", "bc", "d"] # Lists are concatenated,
# but the last and first elements are merged # because they are lists.
# Recursively, the last and first elements # of the inner lists are merged
# because they are strings. ["a", ["b", "c"]], [["d"], "e"] => ["a", ["b", "cd"],
# "e"] # Non-overlapping object fields are combined. `"a": "1"`, `"b": "2"` => `
# "a": "1", "b": 2"` # Overlapping object fields are merged. `"a": "1"`, `"a": "
# 2"` => `"a": "12"` # Examples of merging objects containing lists of strings. `
# "a": ["1"]`, `"a": ["2"]` => `"a": ["12"]` For a more complete example,
# suppose a streaming SQL query is yielding a result set whose rows contain a
# single string field. The following `PartialResultSet`s might be yielded: ` "
# metadata": ` ... ` "values": ["Hello", "W"] "chunked_value": true "
# resume_token": "Af65..." ` ` "values": ["orl"] "chunked_value": true "
# resume_token": "Bqp2..." ` ` "values": ["d"] "resume_token": "Zx1B..." ` This
# sequence of `PartialResultSet`s encodes two rows, one containing the field
# value `"Hello"`, and a second containing the field value `"World" = "W" + "orl"
# + "d"`.
# Corresponds to the JSON property `values`
# @return [Array<Object>]
attr_accessor :values
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@chunked_value = args[:chunked_value] if args.key?(:chunked_value)
@metadata = args[:metadata] if args.key?(:metadata)
@resume_token = args[:resume_token] if args.key?(:resume_token)
@stats = args[:stats] if args.key?(:stats)
@values = args[:values] if args.key?(:values)
end
end
# Information returned for each partition returned in a PartitionResponse.
class Partition
include Google::Apis::Core::Hashable
# This token can be passed to Read, StreamingRead, ExecuteSql, or
# ExecuteStreamingSql requests to restrict the results to those identified by
# this partition token.
# Corresponds to the JSON property `partitionToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :partition_token
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@partition_token = args[:partition_token] if args.key?(:partition_token)
end
end
# Options for a PartitionQueryRequest and PartitionReadRequest.
class PartitionOptions
include Google::Apis::Core::Hashable
# **Note:** This hint is currently ignored by PartitionQuery and PartitionRead
# requests. The desired maximum number of partitions to return. For example,
# this may be set to the number of workers available. The default for this
# option is currently 10,000. The maximum value is currently 200,000. This is
# only a hint. The actual number of partitions returned may be smaller or larger
# than this maximum count request.
# Corresponds to the JSON property `maxPartitions`
# @return [Fixnum]
attr_accessor :max_partitions
# **Note:** This hint is currently ignored by PartitionQuery and PartitionRead
# requests. The desired data size for each partition generated. The default for
# this option is currently 1 GiB. This is only a hint. The actual size of each
# partition may be smaller or larger than this size request.
# Corresponds to the JSON property `partitionSizeBytes`
# @return [Fixnum]
attr_accessor :partition_size_bytes
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@max_partitions = args[:max_partitions] if args.key?(:max_partitions)
@partition_size_bytes = args[:partition_size_bytes] if args.key?(:partition_size_bytes)
end
end
# The request for PartitionQuery
class PartitionQueryRequest
include Google::Apis::Core::Hashable
# It is not always possible for Cloud Spanner to infer the right SQL type from a
# JSON value. For example, values of type `BYTES` and values of type `STRING`
# both appear in params as JSON strings. In these cases, `param_types` can be
# used to specify the exact SQL type for some or all of the SQL query parameters.
# See the definition of Type for more information about SQL types.
# Corresponds to the JSON property `paramTypes`
# @return [Hash<String,Google::Apis::SpannerV1::Type>]
attr_accessor :param_types
# Parameter names and values that bind to placeholders in the SQL string. A
# parameter placeholder consists of the `@` character followed by the parameter
# name (for example, `@firstName`). Parameter names can contain letters, numbers,
# and underscores. Parameters can appear anywhere that a literal value is
# expected. The same parameter name can be used more than once, for example: `"
# WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL
# statement with unbound parameters.
# Corresponds to the JSON property `params`
# @return [Hash<String,Object>]
attr_accessor :params
# Options for a PartitionQueryRequest and PartitionReadRequest.
# Corresponds to the JSON property `partitionOptions`
# @return [Google::Apis::SpannerV1::PartitionOptions]
attr_accessor :partition_options
# Required. The query request to generate partitions for. The request will fail
# if the query is not root partitionable. The query plan of a root partitionable
# query has a single distributed union operator. A distributed union operator
# conceptually divides one or more tables into multiple splits, remotely
# evaluates a subquery independently on each split, and then unions all results.
# This must not contain DML commands, such as INSERT, UPDATE, or DELETE. Use
# ExecuteStreamingSql with a PartitionedDml transaction for large, partition-
# friendly DML operations.
# Corresponds to the JSON property `sql`
# @return [String]
attr_accessor :sql
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::TransactionSelector]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@param_types = args[:param_types] if args.key?(:param_types)
@params = args[:params] if args.key?(:params)
@partition_options = args[:partition_options] if args.key?(:partition_options)
@sql = args[:sql] if args.key?(:sql)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# The request for PartitionRead
class PartitionReadRequest
include Google::Apis::Core::Hashable
# The columns of table to be returned for each row matching this request.
# Corresponds to the JSON property `columns`
# @return [Array<String>]
attr_accessor :columns
# If non-empty, the name of an index on table. This index is used instead of the
# table primary key when interpreting key_set and sorting result rows. See
# key_set for further information.
# Corresponds to the JSON property `index`
# @return [String]
attr_accessor :index
# `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the
# keys are expected to be in the same table or index. The keys need not be
# sorted in any particular way. If the same key is specified multiple times in
# the set (for example if two ranges, two keys, or a key and a range overlap),
# Cloud Spanner behaves as if the key were only specified once.
# Corresponds to the JSON property `keySet`
# @return [Google::Apis::SpannerV1::KeySet]
attr_accessor :key_set
# Options for a PartitionQueryRequest and PartitionReadRequest.
# Corresponds to the JSON property `partitionOptions`
# @return [Google::Apis::SpannerV1::PartitionOptions]
attr_accessor :partition_options
# Required. The name of the table in the database to be read.
# Corresponds to the JSON property `table`
# @return [String]
attr_accessor :table
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::TransactionSelector]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@columns = args[:columns] if args.key?(:columns)
@index = args[:index] if args.key?(:index)
@key_set = args[:key_set] if args.key?(:key_set)
@partition_options = args[:partition_options] if args.key?(:partition_options)
@table = args[:table] if args.key?(:table)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# The response for PartitionQuery or PartitionRead
class PartitionResponse
include Google::Apis::Core::Hashable
# Partitions created by this request.
# Corresponds to the JSON property `partitions`
# @return [Array<Google::Apis::SpannerV1::Partition>]
attr_accessor :partitions
# A transaction.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::Transaction]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@partitions = args[:partitions] if args.key?(:partitions)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# Message type to initiate a Partitioned DML transaction.
class PartitionedDml
include Google::Apis::Core::Hashable
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
end
end
# Node information for nodes appearing in a QueryPlan.plan_nodes.
class PlanNode
include Google::Apis::Core::Hashable
# List of child node `index`es and their relationship to this parent.
# Corresponds to the JSON property `childLinks`
# @return [Array<Google::Apis::SpannerV1::ChildLink>]
attr_accessor :child_links
# The display name for the node.
# Corresponds to the JSON property `displayName`
# @return [String]
attr_accessor :display_name
# The execution statistics associated with the node, contained in a group of key-
# value pairs. Only present if the plan was returned as a result of a profile
# query. For example, number of executions, number of rows/time per execution
# etc.
# Corresponds to the JSON property `executionStats`
# @return [Hash<String,Object>]
attr_accessor :execution_stats
# The `PlanNode`'s index in node list.
# Corresponds to the JSON property `index`
# @return [Fixnum]
attr_accessor :index
# Used to determine the type of node. May be needed for visualizing different
# kinds of nodes differently. For example, If the node is a SCALAR node, it will
# have a condensed representation which can be used to directly embed a
# description of the node in its parent.
# Corresponds to the JSON property `kind`
# @return [String]
attr_accessor :kind
# Attributes relevant to the node contained in a group of key-value pairs. For
# example, a Parameter Reference node could have the following information in
# its metadata: ` "parameter_reference": "param1", "parameter_type": "array" `
# Corresponds to the JSON property `metadata`
# @return [Hash<String,Object>]
attr_accessor :metadata
# Condensed representation of a node and its subtree. Only present for `SCALAR`
# PlanNode(s).
# Corresponds to the JSON property `shortRepresentation`
# @return [Google::Apis::SpannerV1::ShortRepresentation]
attr_accessor :short_representation
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@child_links = args[:child_links] if args.key?(:child_links)
@display_name = args[:display_name] if args.key?(:display_name)
@execution_stats = args[:execution_stats] if args.key?(:execution_stats)
@index = args[:index] if args.key?(:index)
@kind = args[:kind] if args.key?(:kind)
@metadata = args[:metadata] if args.key?(:metadata)
@short_representation = args[:short_representation] if args.key?(:short_representation)
end
end
# An Identity and Access Management (IAM) policy, which specifies access
# controls for Google Cloud resources. A `Policy` is a collection of `bindings`.
# A `binding` binds one or more `members` to a single `role`. Members can be
# user accounts, service accounts, Google groups, and domains (such as G Suite).
# A `role` is a named list of permissions; each `role` can be an IAM predefined
# role or a user-created custom role. For some types of Google Cloud resources,
# a `binding` can also specify a `condition`, which is a logical expression that
# allows access to a resource only if the expression evaluates to `true`. A
# condition can add constraints based on attributes of the request, the resource,
# or both. To learn which resources support conditions in their IAM policies,
# see the [IAM documentation](https://cloud.google.com/iam/help/conditions/
# resource-policies). **JSON example:** ` "bindings": [ ` "role": "roles/
# resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "
# group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@
# appspot.gserviceaccount.com" ] `, ` "role": "roles/resourcemanager.
# organizationViewer", "members": [ "user:eve@example.com" ], "condition": ` "
# title": "expirable access", "description": "Does not grant access after Sep
# 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", `
# ` ], "etag": "BwWWja0YfJA=", "version": 3 ` **YAML example:** bindings: -
# members: - user:mike@example.com - group:admins@example.com - domain:google.
# com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/
# resourcemanager.organizationAdmin - members: - user:eve@example.com role:
# roles/resourcemanager.organizationViewer condition: title: expirable access
# description: Does not grant access after Sep 2020 expression: request.time <
# timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a
# description of IAM and its features, see the [IAM documentation](https://cloud.
# google.com/iam/docs/).
class Policy
include Google::Apis::Core::Hashable
# Associates a list of `members` to a `role`. Optionally, may specify a `
# condition` that determines how and when the `bindings` are applied. Each of
# the `bindings` must contain at least one member.
# Corresponds to the JSON property `bindings`
# @return [Array<Google::Apis::SpannerV1::Binding>]
attr_accessor :bindings
# `etag` is used for optimistic concurrency control as a way to help prevent
# simultaneous updates of a policy from overwriting each other. It is strongly
# suggested that systems make use of the `etag` in the read-modify-write cycle
# to perform policy updates in order to avoid race conditions: An `etag` is
# returned in the response to `getIamPolicy`, and systems are expected to put
# that etag in the request to `setIamPolicy` to ensure that their change will be
# applied to the same version of the policy. **Important:** If you use IAM
# Conditions, you must include the `etag` field whenever you call `setIamPolicy`.
# If you omit this field, then IAM allows you to overwrite a version `3` policy
# with a version `1` policy, and all of the conditions in the version `3` policy
# are lost.
# Corresponds to the JSON property `etag`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :etag
# Specifies the format of the policy. Valid values are `0`, `1`, and `3`.
# Requests that specify an invalid value are rejected. Any operation that
# affects conditional role bindings must specify version `3`. This requirement
# applies to the following operations: * Getting a policy that includes a
# conditional role binding * Adding a conditional role binding to a policy *
# Changing a conditional role binding in a policy * Removing any role binding,
# with or without a condition, from a policy that includes conditions **
# Important:** If you use IAM Conditions, you must include the `etag` field
# whenever you call `setIamPolicy`. If you omit this field, then IAM allows you
# to overwrite a version `3` policy with a version `1` policy, and all of the
# conditions in the version `3` policy are lost. If a policy does not include
# any conditions, operations on that policy may specify any valid version or
# leave the field unset. To learn which resources support conditions in their
# IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/
# conditions/resource-policies).
# Corresponds to the JSON property `version`
# @return [Fixnum]
attr_accessor :version
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@bindings = args[:bindings] if args.key?(:bindings)
@etag = args[:etag] if args.key?(:etag)
@version = args[:version] if args.key?(:version)
end
end
# Query optimizer configuration.
class QueryOptions
include Google::Apis::Core::Hashable
# An option to control the selection of optimizer version. This parameter allows
# individual queries to pick different query optimizer versions. Specifying "
# latest" as a value instructs Cloud Spanner to use the latest supported query
# optimizer version. If not specified, Cloud Spanner uses optimizer version set
# at the database level options. Any other positive integer (from the list of
# supported optimizer versions) overrides the default optimizer version for
# query execution. The list of supported optimizer versions can be queried from
# SPANNER_SYS.SUPPORTED_OPTIMIZER_VERSIONS. Executing a SQL statement with an
# invalid optimizer version will fail with a syntax error (`INVALID_ARGUMENT`)
# status. See https://cloud.google.com/spanner/docs/query-optimizer/manage-query-
# optimizer for more information on managing the query optimizer. The `
# optimizer_version` statement hint has precedence over this setting.
# Corresponds to the JSON property `optimizerVersion`
# @return [String]
attr_accessor :optimizer_version
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@optimizer_version = args[:optimizer_version] if args.key?(:optimizer_version)
end
end
# Contains an ordered list of nodes appearing in the query plan.
class QueryPlan
include Google::Apis::Core::Hashable
# The nodes in the query plan. Plan nodes are returned in pre-order starting
# with the plan root. Each PlanNode's `id` corresponds to its index in `
# plan_nodes`.
# Corresponds to the JSON property `planNodes`
# @return [Array<Google::Apis::SpannerV1::PlanNode>]
attr_accessor :plan_nodes
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@plan_nodes = args[:plan_nodes] if args.key?(:plan_nodes)
end
end
# Message type to initiate a read-only transaction.
class ReadOnly
include Google::Apis::Core::Hashable
# Executes all reads at a timestamp that is `exact_staleness` old. The timestamp
# is chosen soon after the read is started. Guarantees that all writes that have
# committed more than the specified number of seconds ago are visible. Because
# Cloud Spanner chooses the exact timestamp, this mode works even if the client'
# s local clock is substantially skewed from Cloud Spanner commit timestamps.
# Useful for reading at nearby replicas without the distributed timestamp
# negotiation overhead of `max_staleness`.
# Corresponds to the JSON property `exactStaleness`
# @return [String]
attr_accessor :exact_staleness
# Read data at a timestamp >= `NOW - max_staleness` seconds. Guarantees that all
# writes that have committed more than the specified number of seconds ago are
# visible. Because Cloud Spanner chooses the exact timestamp, this mode works
# even if the client's local clock is substantially skewed from Cloud Spanner
# commit timestamps. Useful for reading the freshest data available at a nearby
# replica, while bounding the possible staleness if the local replica has fallen
# behind. Note that this option can only be used in single-use transactions.
# Corresponds to the JSON property `maxStaleness`
# @return [String]
attr_accessor :max_staleness
# Executes all reads at a timestamp >= `min_read_timestamp`. This is useful for
# requesting fresher data than some previous read, or data that is fresh enough
# to observe the effects of some previously committed transaction whose
# timestamp is known. Note that this option can only be used in single-use
# transactions. A timestamp in RFC3339 UTC \"Zulu\" format, accurate to
# nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`.
# Corresponds to the JSON property `minReadTimestamp`
# @return [String]
attr_accessor :min_read_timestamp
# Executes all reads at the given timestamp. Unlike other modes, reads at a
# specific timestamp are repeatable; the same read at the same timestamp always
# returns the same data. If the timestamp is in the future, the read will block
# until the specified timestamp, modulo the read's deadline. Useful for large
# scale consistent reads such as mapreduces, or for coordinating many reads
# against a consistent snapshot of the data. A timestamp in RFC3339 UTC \"Zulu\"
# format, accurate to nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`.
# Corresponds to the JSON property `readTimestamp`
# @return [String]
attr_accessor :read_timestamp
# If true, the Cloud Spanner-selected read timestamp is included in the
# Transaction message that describes the transaction.
# Corresponds to the JSON property `returnReadTimestamp`
# @return [Boolean]
attr_accessor :return_read_timestamp
alias_method :return_read_timestamp?, :return_read_timestamp
# Read at a timestamp where all previously committed transactions are visible.
# Corresponds to the JSON property `strong`
# @return [Boolean]
attr_accessor :strong
alias_method :strong?, :strong
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@exact_staleness = args[:exact_staleness] if args.key?(:exact_staleness)
@max_staleness = args[:max_staleness] if args.key?(:max_staleness)
@min_read_timestamp = args[:min_read_timestamp] if args.key?(:min_read_timestamp)
@read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp)
@return_read_timestamp = args[:return_read_timestamp] if args.key?(:return_read_timestamp)
@strong = args[:strong] if args.key?(:strong)
end
end
# The request for Read and StreamingRead.
class ReadRequest
include Google::Apis::Core::Hashable
# Required. The columns of table to be returned for each row matching this
# request.
# Corresponds to the JSON property `columns`
# @return [Array<String>]
attr_accessor :columns
# If non-empty, the name of an index on table. This index is used instead of the
# table primary key when interpreting key_set and sorting result rows. See
# key_set for further information.
# Corresponds to the JSON property `index`
# @return [String]
attr_accessor :index
# `KeySet` defines a collection of Cloud Spanner keys and/or key ranges. All the
# keys are expected to be in the same table or index. The keys need not be
# sorted in any particular way. If the same key is specified multiple times in
# the set (for example if two ranges, two keys, or a key and a range overlap),
# Cloud Spanner behaves as if the key were only specified once.
# Corresponds to the JSON property `keySet`
# @return [Google::Apis::SpannerV1::KeySet]
attr_accessor :key_set
# If greater than zero, only the first `limit` rows are yielded. If `limit` is
# zero, the default is no limit. A limit cannot be specified if `partition_token`
# is set.
# Corresponds to the JSON property `limit`
# @return [Fixnum]
attr_accessor :limit
# If present, results will be restricted to the specified partition previously
# created using PartitionRead(). There must be an exact match for the values of
# fields common to this message and the PartitionReadRequest message used to
# create this partition_token.
# Corresponds to the JSON property `partitionToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :partition_token
# If this request is resuming a previously interrupted read, `resume_token`
# should be copied from the last PartialResultSet yielded before the
# interruption. Doing this enables the new read to resume where the last read
# left off. The rest of the request parameters must exactly match the request
# that yielded this token.
# Corresponds to the JSON property `resumeToken`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :resume_token
# Required. The name of the table in the database to be read.
# Corresponds to the JSON property `table`
# @return [String]
attr_accessor :table
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::TransactionSelector]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@columns = args[:columns] if args.key?(:columns)
@index = args[:index] if args.key?(:index)
@key_set = args[:key_set] if args.key?(:key_set)
@limit = args[:limit] if args.key?(:limit)
@partition_token = args[:partition_token] if args.key?(:partition_token)
@resume_token = args[:resume_token] if args.key?(:resume_token)
@table = args[:table] if args.key?(:table)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# Message type to initiate a read-write transaction. Currently this transaction
# type has no options.
class ReadWrite
include Google::Apis::Core::Hashable
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
end
end
#
class ReplicaInfo
include Google::Apis::Core::Hashable
# If true, this location is designated as the default leader location where
# leader replicas are placed. See the [region types documentation](https://cloud.
# google.com/spanner/docs/instances#region_types) for more details.
# Corresponds to the JSON property `defaultLeaderLocation`
# @return [Boolean]
attr_accessor :default_leader_location
alias_method :default_leader_location?, :default_leader_location
# The location of the serving resources, e.g. "us-central1".
# Corresponds to the JSON property `location`
# @return [String]
attr_accessor :location
# The type of replica.
# Corresponds to the JSON property `type`
# @return [String]
attr_accessor :type
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@default_leader_location = args[:default_leader_location] if args.key?(:default_leader_location)
@location = args[:location] if args.key?(:location)
@type = args[:type] if args.key?(:type)
end
end
# Metadata type for the long-running operation returned by RestoreDatabase.
class RestoreDatabaseMetadata
include Google::Apis::Core::Hashable
# Information about a backup.
# Corresponds to the JSON property `backupInfo`
# @return [Google::Apis::SpannerV1::BackupInfo]
attr_accessor :backup_info
# The time at which cancellation of this operation was received. Operations.
# CancelOperation starts asynchronous cancellation on a long-running operation.
# The server makes a best effort to cancel the operation, but success is not
# guaranteed. Clients can use Operations.GetOperation or other methods to check
# whether the cancellation succeeded or whether the operation completed despite
# cancellation. On successful cancellation, the operation is not deleted;
# instead, it becomes an operation with an Operation.error value with a google.
# rpc.Status.code of 1, corresponding to `Code.CANCELLED`.
# Corresponds to the JSON property `cancelTime`
# @return [String]
attr_accessor :cancel_time
# Name of the database being created and restored to.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
# If exists, the name of the long-running operation that will be used to track
# the post-restore optimization process to optimize the performance of the
# restored database, and remove the dependency on the restore source. The name
# is of the form `projects//instances//databases//operations/` where the is the
# name of database being created and restored to. The metadata type of the long-
# running operation is OptimizeRestoredDatabaseMetadata. This long-running
# operation will be automatically created by the system after the
# RestoreDatabase long-running operation completes successfully. This operation
# will not be created if the restore was not successful.
# Corresponds to the JSON property `optimizeDatabaseOperationName`
# @return [String]
attr_accessor :optimize_database_operation_name
# Encapsulates progress related information for a Cloud Spanner long running
# operation.
# Corresponds to the JSON property `progress`
# @return [Google::Apis::SpannerV1::OperationProgress]
attr_accessor :progress
# The type of the restore source.
# Corresponds to the JSON property `sourceType`
# @return [String]
attr_accessor :source_type
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@backup_info = args[:backup_info] if args.key?(:backup_info)
@cancel_time = args[:cancel_time] if args.key?(:cancel_time)
@name = args[:name] if args.key?(:name)
@optimize_database_operation_name = args[:optimize_database_operation_name] if args.key?(:optimize_database_operation_name)
@progress = args[:progress] if args.key?(:progress)
@source_type = args[:source_type] if args.key?(:source_type)
end
end
# The request for RestoreDatabase.
class RestoreDatabaseRequest
include Google::Apis::Core::Hashable
# Name of the backup from which to restore. Values are of the form `projects//
# instances//backups/`.
# Corresponds to the JSON property `backup`
# @return [String]
attr_accessor :backup
# Required. The id of the database to create and restore to. This database must
# not already exist. The `database_id` appended to `parent` forms the full
# database name of the form `projects//instances//databases/`.
# Corresponds to the JSON property `databaseId`
# @return [String]
attr_accessor :database_id
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@backup = args[:backup] if args.key?(:backup)
@database_id = args[:database_id] if args.key?(:database_id)
end
end
# Information about the database restore.
class RestoreInfo
include Google::Apis::Core::Hashable
# Information about a backup.
# Corresponds to the JSON property `backupInfo`
# @return [Google::Apis::SpannerV1::BackupInfo]
attr_accessor :backup_info
# The type of the restore source.
# Corresponds to the JSON property `sourceType`
# @return [String]
attr_accessor :source_type
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@backup_info = args[:backup_info] if args.key?(:backup_info)
@source_type = args[:source_type] if args.key?(:source_type)
end
end
# Results from Read or ExecuteSql.
class ResultSet
include Google::Apis::Core::Hashable
# Metadata about a ResultSet or PartialResultSet.
# Corresponds to the JSON property `metadata`
# @return [Google::Apis::SpannerV1::ResultSetMetadata]
attr_accessor :metadata
# Each element in `rows` is a row whose format is defined by metadata.row_type.
# The ith element in each row matches the ith field in metadata.row_type.
# Elements are encoded based on type as described here.
# Corresponds to the JSON property `rows`
# @return [Array<Array<Object>>]
attr_accessor :rows
# Additional statistics about a ResultSet or PartialResultSet.
# Corresponds to the JSON property `stats`
# @return [Google::Apis::SpannerV1::ResultSetStats]
attr_accessor :stats
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@metadata = args[:metadata] if args.key?(:metadata)
@rows = args[:rows] if args.key?(:rows)
@stats = args[:stats] if args.key?(:stats)
end
end
# Metadata about a ResultSet or PartialResultSet.
class ResultSetMetadata
include Google::Apis::Core::Hashable
# `StructType` defines the fields of a STRUCT type.
# Corresponds to the JSON property `rowType`
# @return [Google::Apis::SpannerV1::StructType]
attr_accessor :row_type
# A transaction.
# Corresponds to the JSON property `transaction`
# @return [Google::Apis::SpannerV1::Transaction]
attr_accessor :transaction
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@row_type = args[:row_type] if args.key?(:row_type)
@transaction = args[:transaction] if args.key?(:transaction)
end
end
# Additional statistics about a ResultSet or PartialResultSet.
class ResultSetStats
include Google::Apis::Core::Hashable
# Contains an ordered list of nodes appearing in the query plan.
# Corresponds to the JSON property `queryPlan`
# @return [Google::Apis::SpannerV1::QueryPlan]
attr_accessor :query_plan
# Aggregated statistics from the execution of the query. Only present when the
# query is profiled. For example, a query could return the statistics as follows:
# ` "rows_returned": "3", "elapsed_time": "1.22 secs", "cpu_time": "1.19 secs" `
# Corresponds to the JSON property `queryStats`
# @return [Hash<String,Object>]
attr_accessor :query_stats
# Standard DML returns an exact count of rows that were modified.
# Corresponds to the JSON property `rowCountExact`
# @return [Fixnum]
attr_accessor :row_count_exact
# Partitioned DML does not offer exactly-once semantics, so it returns a lower
# bound of the rows modified.
# Corresponds to the JSON property `rowCountLowerBound`
# @return [Fixnum]
attr_accessor :row_count_lower_bound
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@query_plan = args[:query_plan] if args.key?(:query_plan)
@query_stats = args[:query_stats] if args.key?(:query_stats)
@row_count_exact = args[:row_count_exact] if args.key?(:row_count_exact)
@row_count_lower_bound = args[:row_count_lower_bound] if args.key?(:row_count_lower_bound)
end
end
# The request for Rollback.
class RollbackRequest
include Google::Apis::Core::Hashable
# Required. The transaction to roll back.
# Corresponds to the JSON property `transactionId`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :transaction_id
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@transaction_id = args[:transaction_id] if args.key?(:transaction_id)
end
end
# A session in the Cloud Spanner API.
class Session
include Google::Apis::Core::Hashable
# Output only. The approximate timestamp when the session is last used. It is
# typically earlier than the actual last use time.
# Corresponds to the JSON property `approximateLastUseTime`
# @return [String]
attr_accessor :approximate_last_use_time
# Output only. The timestamp when the session is created.
# Corresponds to the JSON property `createTime`
# @return [String]
attr_accessor :create_time
# The labels for the session. * Label keys must be between 1 and 63 characters
# long and must conform to the following regular expression: `[a-z]([-a-z0-9]*[a-
# z0-9])?`. * Label values must be between 0 and 63 characters long and must
# conform to the regular expression `([a-z]([-a-z0-9]*[a-z0-9])?)?`. * No more
# than 64 labels can be associated with a given session. See https://goo.gl/
# xmQnxf for more information on and examples of labels.
# Corresponds to the JSON property `labels`
# @return [Hash<String,String>]
attr_accessor :labels
# Output only. The name of the session. This is always system-assigned.
# Corresponds to the JSON property `name`
# @return [String]
attr_accessor :name
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@approximate_last_use_time = args[:approximate_last_use_time] if args.key?(:approximate_last_use_time)
@create_time = args[:create_time] if args.key?(:create_time)
@labels = args[:labels] if args.key?(:labels)
@name = args[:name] if args.key?(:name)
end
end
# Request message for `SetIamPolicy` method.
class SetIamPolicyRequest
include Google::Apis::Core::Hashable
# An Identity and Access Management (IAM) policy, which specifies access
# controls for Google Cloud resources. A `Policy` is a collection of `bindings`.
# A `binding` binds one or more `members` to a single `role`. Members can be
# user accounts, service accounts, Google groups, and domains (such as G Suite).
# A `role` is a named list of permissions; each `role` can be an IAM predefined
# role or a user-created custom role. For some types of Google Cloud resources,
# a `binding` can also specify a `condition`, which is a logical expression that
# allows access to a resource only if the expression evaluates to `true`. A
# condition can add constraints based on attributes of the request, the resource,
# or both. To learn which resources support conditions in their IAM policies,
# see the [IAM documentation](https://cloud.google.com/iam/help/conditions/
# resource-policies). **JSON example:** ` "bindings": [ ` "role": "roles/
# resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "
# group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@
# appspot.gserviceaccount.com" ] `, ` "role": "roles/resourcemanager.
# organizationViewer", "members": [ "user:eve@example.com" ], "condition": ` "
# title": "expirable access", "description": "Does not grant access after Sep
# 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", `
# ` ], "etag": "BwWWja0YfJA=", "version": 3 ` **YAML example:** bindings: -
# members: - user:mike@example.com - group:admins@example.com - domain:google.
# com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/
# resourcemanager.organizationAdmin - members: - user:eve@example.com role:
# roles/resourcemanager.organizationViewer condition: title: expirable access
# description: Does not grant access after Sep 2020 expression: request.time <
# timestamp('2020-10-01T00:00:00.000Z') - etag: BwWWja0YfJA= - version: 3 For a
# description of IAM and its features, see the [IAM documentation](https://cloud.
# google.com/iam/docs/).
# Corresponds to the JSON property `policy`
# @return [Google::Apis::SpannerV1::Policy]
attr_accessor :policy
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@policy = args[:policy] if args.key?(:policy)
end
end
# Condensed representation of a node and its subtree. Only present for `SCALAR`
# PlanNode(s).
class ShortRepresentation
include Google::Apis::Core::Hashable
# A string representation of the expression subtree rooted at this node.
# Corresponds to the JSON property `description`
# @return [String]
attr_accessor :description
# A mapping of (subquery variable name) -> (subquery node id) for cases where
# the `description` string of this node references a `SCALAR` subquery contained
# in the expression subtree rooted at this node. The referenced `SCALAR`
# subquery may not necessarily be a direct child of this node.
# Corresponds to the JSON property `subqueries`
# @return [Hash<String,Fixnum>]
attr_accessor :subqueries
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@description = args[:description] if args.key?(:description)
@subqueries = args[:subqueries] if args.key?(:subqueries)
end
end
# A single DML statement.
class Statement
include Google::Apis::Core::Hashable
# It is not always possible for Cloud Spanner to infer the right SQL type from a
# JSON value. For example, values of type `BYTES` and values of type `STRING`
# both appear in params as JSON strings. In these cases, `param_types` can be
# used to specify the exact SQL type for some or all of the SQL statement
# parameters. See the definition of Type for more information about SQL types.
# Corresponds to the JSON property `paramTypes`
# @return [Hash<String,Google::Apis::SpannerV1::Type>]
attr_accessor :param_types
# Parameter names and values that bind to placeholders in the DML string. A
# parameter placeholder consists of the `@` character followed by the parameter
# name (for example, `@firstName`). Parameter names can contain letters, numbers,
# and underscores. Parameters can appear anywhere that a literal value is
# expected. The same parameter name can be used more than once, for example: `"
# WHERE id > @msg_id AND id < @msg_id + 100"` It is an error to execute a SQL
# statement with unbound parameters.
# Corresponds to the JSON property `params`
# @return [Hash<String,Object>]
attr_accessor :params
# Required. The DML string.
# Corresponds to the JSON property `sql`
# @return [String]
attr_accessor :sql
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@param_types = args[:param_types] if args.key?(:param_types)
@params = args[:params] if args.key?(:params)
@sql = args[:sql] if args.key?(:sql)
end
end
# The `Status` type defines a logical error model that is suitable for different
# programming environments, including REST APIs and RPC APIs. It is used by [
# gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
# data: error code, error message, and error details. You can find out more
# about this error model and how to work with it in the [API Design Guide](https:
# //cloud.google.com/apis/design/errors).
class Status
include Google::Apis::Core::Hashable
# The status code, which should be an enum value of google.rpc.Code.
# Corresponds to the JSON property `code`
# @return [Fixnum]
attr_accessor :code
# A list of messages that carry the error details. There is a common set of
# message types for APIs to use.
# Corresponds to the JSON property `details`
# @return [Array<Hash<String,Object>>]
attr_accessor :details
# A developer-facing error message, which should be in English. Any user-facing
# error message should be localized and sent in the google.rpc.Status.details
# field, or localized by the client.
# Corresponds to the JSON property `message`
# @return [String]
attr_accessor :message
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@code = args[:code] if args.key?(:code)
@details = args[:details] if args.key?(:details)
@message = args[:message] if args.key?(:message)
end
end
# `StructType` defines the fields of a STRUCT type.
class StructType
include Google::Apis::Core::Hashable
# The list of fields that make up this struct. Order is significant, because
# values of this struct type are represented as lists, where the order of field
# values matches the order of fields in the StructType. In turn, the order of
# fields matches the order of columns in a read request, or the order of fields
# in the `SELECT` clause of a query.
# Corresponds to the JSON property `fields`
# @return [Array<Google::Apis::SpannerV1::Field>]
attr_accessor :fields
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@fields = args[:fields] if args.key?(:fields)
end
end
# Request message for `TestIamPermissions` method.
class TestIamPermissionsRequest
include Google::Apis::Core::Hashable
# REQUIRED: The set of permissions to check for 'resource'. Permissions with
# wildcards (such as '*', 'spanner.*', 'spanner.instances.*') are not allowed.
# Corresponds to the JSON property `permissions`
# @return [Array<String>]
attr_accessor :permissions
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@permissions = args[:permissions] if args.key?(:permissions)
end
end
# Response message for `TestIamPermissions` method.
class TestIamPermissionsResponse
include Google::Apis::Core::Hashable
# A subset of `TestPermissionsRequest.permissions` that the caller is allowed.
# Corresponds to the JSON property `permissions`
# @return [Array<String>]
attr_accessor :permissions
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@permissions = args[:permissions] if args.key?(:permissions)
end
end
# A transaction.
class Transaction
include Google::Apis::Core::Hashable
# `id` may be used to identify the transaction in subsequent Read, ExecuteSql,
# Commit, or Rollback calls. Single-use read-only transactions do not have IDs,
# because single-use transactions do not support multiple requests.
# Corresponds to the JSON property `id`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :id
# For snapshot read-only transactions, the read timestamp chosen for the
# transaction. Not returned by default: see TransactionOptions.ReadOnly.
# return_read_timestamp. A timestamp in RFC3339 UTC \"Zulu\" format, accurate to
# nanoseconds. Example: `"2014-10-02T15:01:23.045123456Z"`.
# Corresponds to the JSON property `readTimestamp`
# @return [String]
attr_accessor :read_timestamp
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@id = args[:id] if args.key?(:id)
@read_timestamp = args[:read_timestamp] if args.key?(:read_timestamp)
end
end
# # Transactions Each session can have at most one active transaction at a time (
# note that standalone reads and queries use a transaction internally and do
# count towards the one transaction limit). After the active transaction is
# completed, the session can immediately be re-used for the next transaction. It
# is not necessary to create a new session for each transaction. # Transaction
# Modes Cloud Spanner supports three transaction modes: 1. Locking read-write.
# This type of transaction is the only way to write data into Cloud Spanner.
# These transactions rely on pessimistic locking and, if necessary, two-phase
# commit. Locking read-write transactions may abort, requiring the application
# to retry. 2. Snapshot read-only. This transaction type provides guaranteed
# consistency across several reads, but does not allow writes. Snapshot read-
# only transactions can be configured to read at timestamps in the past.
# Snapshot read-only transactions do not need to be committed. 3. Partitioned
# DML. This type of transaction is used to execute a single Partitioned DML
# statement. Partitioned DML partitions the key space and runs the DML statement
# over each partition in parallel using separate, internal transactions that
# commit independently. Partitioned DML transactions do not need to be committed.
# For transactions that only read, snapshot read-only transactions provide
# simpler semantics and are almost always faster. In particular, read-only
# transactions do not take locks, so they do not conflict with read-write
# transactions. As a consequence of not taking locks, they also do not abort, so
# retry loops are not needed. Transactions may only read/write data in a single
# database. They may, however, read/write data in different tables within that
# database. ## Locking Read-Write Transactions Locking transactions may be used
# to atomically read-modify-write data anywhere in a database. This type of
# transaction is externally consistent. Clients should attempt to minimize the
# amount of time a transaction is active. Faster transactions commit with higher
# probability and cause less contention. Cloud Spanner attempts to keep read
# locks active as long as the transaction continues to do reads, and the
# transaction has not been terminated by Commit or Rollback. Long periods of
# inactivity at the client may cause Cloud Spanner to release a transaction's
# locks and abort it. Conceptually, a read-write transaction consists of zero or
# more reads or SQL statements followed by Commit. At any time before Commit,
# the client can send a Rollback request to abort the transaction. ### Semantics
# Cloud Spanner can commit the transaction if all read locks it acquired are
# still valid at commit time, and it is able to acquire write locks for all
# writes. Cloud Spanner can abort the transaction for any reason. If a commit
# attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has
# not modified any user data in Cloud Spanner. Unless the transaction commits,
# Cloud Spanner makes no guarantees about how long the transaction's locks were
# held for. It is an error to use Cloud Spanner locks for any sort of mutual
# exclusion other than between Cloud Spanner transactions themselves. ###
# Retrying Aborted Transactions When a transaction aborts, the application can
# choose to retry the whole transaction again. To maximize the chances of
# successfully committing the retry, the client should execute the retry in the
# same session as the original attempt. The original session's lock priority
# increases with each consecutive abort, meaning that each attempt has a
# slightly better chance of success than the previous. Under some circumstances (
# e.g., many transactions attempting to modify the same row(s)), a transaction
# can abort many times in a short period before successfully committing. Thus,
# it is not a good idea to cap the number of retries a transaction can attempt;
# instead, it is better to limit the total amount of wall time spent retrying. ##
# # Idle Transactions A transaction is considered idle if it has no outstanding
# reads or SQL queries and has not started a read or SQL query within the last
# 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don'
# t hold on to locks indefinitely. In that case, the commit will fail with error
# `ABORTED`. If this behavior is undesirable, periodically executing a simple
# SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from
# becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only
# transactions provides a simpler method than locking read-write transactions
# for doing several consistent reads. However, this type of transaction does not
# support writes. Snapshot transactions do not take locks. Instead, they work by
# choosing a Cloud Spanner timestamp, then executing all reads at that timestamp.
# Since they do not acquire locks, they do not block concurrent read-write
# transactions. Unlike locking read-write transactions, snapshot read-only
# transactions never abort. They can fail if the chosen read timestamp is
# garbage collected; however, the default garbage collection policy is generous
# enough that most applications do not need to worry about this in practice.
# Snapshot read-only transactions do not need to call Commit or Rollback (and in
# fact are not permitted to do so). To execute a snapshot transaction, the
# client specifies a timestamp bound, which tells Cloud Spanner how to choose a
# read timestamp. The types of timestamp bound are: - Strong (the default). -
# Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read
# is geographically distributed, stale read-only transactions can execute more
# quickly than strong or read-write transaction, because they are able to
# execute far from the leader replica. Each type of timestamp bound is discussed
# in detail below. ### Strong Strong reads are guaranteed to see the effects of
# all transactions that have committed before the start of the read. Furthermore,
# all rows yielded by a single read are consistent with each other -- if any
# part of the read observes a transaction, all parts of the read see the
# transaction. Strong reads are not repeatable: two consecutive strong read-only
# transactions might return inconsistent results if there are concurrent writes.
# If consistency across reads is required, the reads should be executed within a
# transaction or at an exact read timestamp. See TransactionOptions.ReadOnly.
# strong. ### Exact Staleness These timestamp bounds execute reads at a user-
# specified timestamp. Reads at a timestamp are guaranteed to see a consistent
# prefix of the global transaction history: they observe modifications done by
# all transactions with a commit timestamp <= the read timestamp, and observe
# none of the modifications done by transactions with a larger commit timestamp.
# They will block until all conflicting transactions that may be assigned commit
# timestamps <= the read timestamp have finished. The timestamp can either be
# expressed as an absolute Cloud Spanner commit timestamp or a staleness
# relative to the current time. These modes do not require a "negotiation phase"
# to pick a timestamp. As a result, they execute slightly faster than the
# equivalent boundedly stale concurrency modes. On the other hand, boundedly
# stale reads usually return fresher results. See TransactionOptions.ReadOnly.
# read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded
# Staleness Bounded staleness modes allow Cloud Spanner to pick the read
# timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses
# the newest timestamp within the staleness bound that allows execution of the
# reads at the closest available replica without blocking. All rows yielded are
# consistent with each other -- if any part of the read observes a transaction,
# all parts of the read see the transaction. Boundedly stale reads are not
# repeatable: two stale reads, even if they use the same staleness bound, can
# execute at different timestamps and thus return inconsistent results.
# Boundedly stale reads execute in two phases: the first phase negotiates a
# timestamp among all replicas needed to serve the read. In the second phase,
# reads are executed at the negotiated timestamp. As a result of the two phase
# execution, bounded staleness reads are usually a little slower than comparable
# exact staleness reads. However, they are typically able to return fresher
# results, and are more likely to execute at the closest replica. Because the
# timestamp negotiation requires up-front knowledge of which rows will be read,
# it can only be used with single-use read-only transactions. See
# TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.
# min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud
# Spanner continuously garbage collects deleted and overwritten data in the
# background to reclaim storage space. This process is known as "version GC". By
# default, version GC reclaims versions after they are one hour old. Because of
# this, Cloud Spanner cannot perform reads at read timestamps more than one hour
# in the past. This restriction also applies to in-progress reads and/or SQL
# queries whose timestamp become too old while executing. Reads and SQL queries
# with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ##
# Partitioned DML Transactions Partitioned DML transactions are used to execute
# DML statements with a different execution strategy that provides different,
# and often better, scalability properties for large, table-wide operations than
# DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP
# workload, should prefer using ReadWrite transactions. Partitioned DML
# partitions the keyspace and runs the DML statement on each partition in
# separate, internal transactions. These transactions commit automatically when
# complete, and run independently from one another. To reduce lock contention,
# this execution strategy only acquires read locks on rows that match the WHERE
# clause of the statement. Additionally, the smaller per-partition transactions
# hold locks for less time. That said, Partitioned DML is not a drop-in
# replacement for standard DML used in ReadWrite transactions. - The DML
# statement must be fully-partitionable. Specifically, the statement must be
# expressible as the union of many statements which each access only a single
# row of the table. - The statement is not applied atomically to all rows of the
# table. Rather, the statement is applied atomically to partitions of the table,
# in independent transactions. Secondary index rows are updated atomically with
# the base table rows. - Partitioned DML does not guarantee exactly-once
# execution semantics against a partition. The statement will be applied at
# least once to each partition. It is strongly recommended that the DML
# statement should be idempotent to avoid unexpected results. For instance, it
# is potentially dangerous to run a statement such as `UPDATE table SET column =
# column + 1` as it could be run multiple times against some rows. - The
# partitions are committed automatically - there is no support for Commit or
# Rollback. If the call returns an error, or if the client issuing the
# ExecuteSql call dies, it is possible that some rows had the statement executed
# on them successfully. It is also possible that statement was never executed
# against other rows. - Partitioned DML transactions may only contain the
# execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. -
# If any error is encountered during the execution of the partitioned DML
# operation (for instance, a UNIQUE INDEX violation, division by zero, or a
# value that cannot be stored due to schema constraints), then the operation is
# stopped at that point and an error is returned. It is possible that at this
# point, some partitions have been committed (or even committed multiple times),
# and other partitions have not been run at all. Given the above, Partitioned
# DML is good fit for large, database-wide, operations that are idempotent, such
# as deleting old rows from a very large table.
class TransactionOptions
include Google::Apis::Core::Hashable
# Message type to initiate a Partitioned DML transaction.
# Corresponds to the JSON property `partitionedDml`
# @return [Google::Apis::SpannerV1::PartitionedDml]
attr_accessor :partitioned_dml
# Message type to initiate a read-only transaction.
# Corresponds to the JSON property `readOnly`
# @return [Google::Apis::SpannerV1::ReadOnly]
attr_accessor :read_only
# Message type to initiate a read-write transaction. Currently this transaction
# type has no options.
# Corresponds to the JSON property `readWrite`
# @return [Google::Apis::SpannerV1::ReadWrite]
attr_accessor :read_write
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@partitioned_dml = args[:partitioned_dml] if args.key?(:partitioned_dml)
@read_only = args[:read_only] if args.key?(:read_only)
@read_write = args[:read_write] if args.key?(:read_write)
end
end
# This message is used to select the transaction in which a Read or ExecuteSql
# call runs. See TransactionOptions for more information about transactions.
class TransactionSelector
include Google::Apis::Core::Hashable
# # Transactions Each session can have at most one active transaction at a time (
# note that standalone reads and queries use a transaction internally and do
# count towards the one transaction limit). After the active transaction is
# completed, the session can immediately be re-used for the next transaction. It
# is not necessary to create a new session for each transaction. # Transaction
# Modes Cloud Spanner supports three transaction modes: 1. Locking read-write.
# This type of transaction is the only way to write data into Cloud Spanner.
# These transactions rely on pessimistic locking and, if necessary, two-phase
# commit. Locking read-write transactions may abort, requiring the application
# to retry. 2. Snapshot read-only. This transaction type provides guaranteed
# consistency across several reads, but does not allow writes. Snapshot read-
# only transactions can be configured to read at timestamps in the past.
# Snapshot read-only transactions do not need to be committed. 3. Partitioned
# DML. This type of transaction is used to execute a single Partitioned DML
# statement. Partitioned DML partitions the key space and runs the DML statement
# over each partition in parallel using separate, internal transactions that
# commit independently. Partitioned DML transactions do not need to be committed.
# For transactions that only read, snapshot read-only transactions provide
# simpler semantics and are almost always faster. In particular, read-only
# transactions do not take locks, so they do not conflict with read-write
# transactions. As a consequence of not taking locks, they also do not abort, so
# retry loops are not needed. Transactions may only read/write data in a single
# database. They may, however, read/write data in different tables within that
# database. ## Locking Read-Write Transactions Locking transactions may be used
# to atomically read-modify-write data anywhere in a database. This type of
# transaction is externally consistent. Clients should attempt to minimize the
# amount of time a transaction is active. Faster transactions commit with higher
# probability and cause less contention. Cloud Spanner attempts to keep read
# locks active as long as the transaction continues to do reads, and the
# transaction has not been terminated by Commit or Rollback. Long periods of
# inactivity at the client may cause Cloud Spanner to release a transaction's
# locks and abort it. Conceptually, a read-write transaction consists of zero or
# more reads or SQL statements followed by Commit. At any time before Commit,
# the client can send a Rollback request to abort the transaction. ### Semantics
# Cloud Spanner can commit the transaction if all read locks it acquired are
# still valid at commit time, and it is able to acquire write locks for all
# writes. Cloud Spanner can abort the transaction for any reason. If a commit
# attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has
# not modified any user data in Cloud Spanner. Unless the transaction commits,
# Cloud Spanner makes no guarantees about how long the transaction's locks were
# held for. It is an error to use Cloud Spanner locks for any sort of mutual
# exclusion other than between Cloud Spanner transactions themselves. ###
# Retrying Aborted Transactions When a transaction aborts, the application can
# choose to retry the whole transaction again. To maximize the chances of
# successfully committing the retry, the client should execute the retry in the
# same session as the original attempt. The original session's lock priority
# increases with each consecutive abort, meaning that each attempt has a
# slightly better chance of success than the previous. Under some circumstances (
# e.g., many transactions attempting to modify the same row(s)), a transaction
# can abort many times in a short period before successfully committing. Thus,
# it is not a good idea to cap the number of retries a transaction can attempt;
# instead, it is better to limit the total amount of wall time spent retrying. ##
# # Idle Transactions A transaction is considered idle if it has no outstanding
# reads or SQL queries and has not started a read or SQL query within the last
# 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don'
# t hold on to locks indefinitely. In that case, the commit will fail with error
# `ABORTED`. If this behavior is undesirable, periodically executing a simple
# SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from
# becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only
# transactions provides a simpler method than locking read-write transactions
# for doing several consistent reads. However, this type of transaction does not
# support writes. Snapshot transactions do not take locks. Instead, they work by
# choosing a Cloud Spanner timestamp, then executing all reads at that timestamp.
# Since they do not acquire locks, they do not block concurrent read-write
# transactions. Unlike locking read-write transactions, snapshot read-only
# transactions never abort. They can fail if the chosen read timestamp is
# garbage collected; however, the default garbage collection policy is generous
# enough that most applications do not need to worry about this in practice.
# Snapshot read-only transactions do not need to call Commit or Rollback (and in
# fact are not permitted to do so). To execute a snapshot transaction, the
# client specifies a timestamp bound, which tells Cloud Spanner how to choose a
# read timestamp. The types of timestamp bound are: - Strong (the default). -
# Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read
# is geographically distributed, stale read-only transactions can execute more
# quickly than strong or read-write transaction, because they are able to
# execute far from the leader replica. Each type of timestamp bound is discussed
# in detail below. ### Strong Strong reads are guaranteed to see the effects of
# all transactions that have committed before the start of the read. Furthermore,
# all rows yielded by a single read are consistent with each other -- if any
# part of the read observes a transaction, all parts of the read see the
# transaction. Strong reads are not repeatable: two consecutive strong read-only
# transactions might return inconsistent results if there are concurrent writes.
# If consistency across reads is required, the reads should be executed within a
# transaction or at an exact read timestamp. See TransactionOptions.ReadOnly.
# strong. ### Exact Staleness These timestamp bounds execute reads at a user-
# specified timestamp. Reads at a timestamp are guaranteed to see a consistent
# prefix of the global transaction history: they observe modifications done by
# all transactions with a commit timestamp <= the read timestamp, and observe
# none of the modifications done by transactions with a larger commit timestamp.
# They will block until all conflicting transactions that may be assigned commit
# timestamps <= the read timestamp have finished. The timestamp can either be
# expressed as an absolute Cloud Spanner commit timestamp or a staleness
# relative to the current time. These modes do not require a "negotiation phase"
# to pick a timestamp. As a result, they execute slightly faster than the
# equivalent boundedly stale concurrency modes. On the other hand, boundedly
# stale reads usually return fresher results. See TransactionOptions.ReadOnly.
# read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded
# Staleness Bounded staleness modes allow Cloud Spanner to pick the read
# timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses
# the newest timestamp within the staleness bound that allows execution of the
# reads at the closest available replica without blocking. All rows yielded are
# consistent with each other -- if any part of the read observes a transaction,
# all parts of the read see the transaction. Boundedly stale reads are not
# repeatable: two stale reads, even if they use the same staleness bound, can
# execute at different timestamps and thus return inconsistent results.
# Boundedly stale reads execute in two phases: the first phase negotiates a
# timestamp among all replicas needed to serve the read. In the second phase,
# reads are executed at the negotiated timestamp. As a result of the two phase
# execution, bounded staleness reads are usually a little slower than comparable
# exact staleness reads. However, they are typically able to return fresher
# results, and are more likely to execute at the closest replica. Because the
# timestamp negotiation requires up-front knowledge of which rows will be read,
# it can only be used with single-use read-only transactions. See
# TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.
# min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud
# Spanner continuously garbage collects deleted and overwritten data in the
# background to reclaim storage space. This process is known as "version GC". By
# default, version GC reclaims versions after they are one hour old. Because of
# this, Cloud Spanner cannot perform reads at read timestamps more than one hour
# in the past. This restriction also applies to in-progress reads and/or SQL
# queries whose timestamp become too old while executing. Reads and SQL queries
# with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ##
# Partitioned DML Transactions Partitioned DML transactions are used to execute
# DML statements with a different execution strategy that provides different,
# and often better, scalability properties for large, table-wide operations than
# DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP
# workload, should prefer using ReadWrite transactions. Partitioned DML
# partitions the keyspace and runs the DML statement on each partition in
# separate, internal transactions. These transactions commit automatically when
# complete, and run independently from one another. To reduce lock contention,
# this execution strategy only acquires read locks on rows that match the WHERE
# clause of the statement. Additionally, the smaller per-partition transactions
# hold locks for less time. That said, Partitioned DML is not a drop-in
# replacement for standard DML used in ReadWrite transactions. - The DML
# statement must be fully-partitionable. Specifically, the statement must be
# expressible as the union of many statements which each access only a single
# row of the table. - The statement is not applied atomically to all rows of the
# table. Rather, the statement is applied atomically to partitions of the table,
# in independent transactions. Secondary index rows are updated atomically with
# the base table rows. - Partitioned DML does not guarantee exactly-once
# execution semantics against a partition. The statement will be applied at
# least once to each partition. It is strongly recommended that the DML
# statement should be idempotent to avoid unexpected results. For instance, it
# is potentially dangerous to run a statement such as `UPDATE table SET column =
# column + 1` as it could be run multiple times against some rows. - The
# partitions are committed automatically - there is no support for Commit or
# Rollback. If the call returns an error, or if the client issuing the
# ExecuteSql call dies, it is possible that some rows had the statement executed
# on them successfully. It is also possible that statement was never executed
# against other rows. - Partitioned DML transactions may only contain the
# execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. -
# If any error is encountered during the execution of the partitioned DML
# operation (for instance, a UNIQUE INDEX violation, division by zero, or a
# value that cannot be stored due to schema constraints), then the operation is
# stopped at that point and an error is returned. It is possible that at this
# point, some partitions have been committed (or even committed multiple times),
# and other partitions have not been run at all. Given the above, Partitioned
# DML is good fit for large, database-wide, operations that are idempotent, such
# as deleting old rows from a very large table.
# Corresponds to the JSON property `begin`
# @return [Google::Apis::SpannerV1::TransactionOptions]
attr_accessor :begin
# Execute the read or SQL query in a previously-started transaction.
# Corresponds to the JSON property `id`
# NOTE: Values are automatically base64 encoded/decoded in the client library.
# @return [String]
attr_accessor :id
# # Transactions Each session can have at most one active transaction at a time (
# note that standalone reads and queries use a transaction internally and do
# count towards the one transaction limit). After the active transaction is
# completed, the session can immediately be re-used for the next transaction. It
# is not necessary to create a new session for each transaction. # Transaction
# Modes Cloud Spanner supports three transaction modes: 1. Locking read-write.
# This type of transaction is the only way to write data into Cloud Spanner.
# These transactions rely on pessimistic locking and, if necessary, two-phase
# commit. Locking read-write transactions may abort, requiring the application
# to retry. 2. Snapshot read-only. This transaction type provides guaranteed
# consistency across several reads, but does not allow writes. Snapshot read-
# only transactions can be configured to read at timestamps in the past.
# Snapshot read-only transactions do not need to be committed. 3. Partitioned
# DML. This type of transaction is used to execute a single Partitioned DML
# statement. Partitioned DML partitions the key space and runs the DML statement
# over each partition in parallel using separate, internal transactions that
# commit independently. Partitioned DML transactions do not need to be committed.
# For transactions that only read, snapshot read-only transactions provide
# simpler semantics and are almost always faster. In particular, read-only
# transactions do not take locks, so they do not conflict with read-write
# transactions. As a consequence of not taking locks, they also do not abort, so
# retry loops are not needed. Transactions may only read/write data in a single
# database. They may, however, read/write data in different tables within that
# database. ## Locking Read-Write Transactions Locking transactions may be used
# to atomically read-modify-write data anywhere in a database. This type of
# transaction is externally consistent. Clients should attempt to minimize the
# amount of time a transaction is active. Faster transactions commit with higher
# probability and cause less contention. Cloud Spanner attempts to keep read
# locks active as long as the transaction continues to do reads, and the
# transaction has not been terminated by Commit or Rollback. Long periods of
# inactivity at the client may cause Cloud Spanner to release a transaction's
# locks and abort it. Conceptually, a read-write transaction consists of zero or
# more reads or SQL statements followed by Commit. At any time before Commit,
# the client can send a Rollback request to abort the transaction. ### Semantics
# Cloud Spanner can commit the transaction if all read locks it acquired are
# still valid at commit time, and it is able to acquire write locks for all
# writes. Cloud Spanner can abort the transaction for any reason. If a commit
# attempt returns `ABORTED`, Cloud Spanner guarantees that the transaction has
# not modified any user data in Cloud Spanner. Unless the transaction commits,
# Cloud Spanner makes no guarantees about how long the transaction's locks were
# held for. It is an error to use Cloud Spanner locks for any sort of mutual
# exclusion other than between Cloud Spanner transactions themselves. ###
# Retrying Aborted Transactions When a transaction aborts, the application can
# choose to retry the whole transaction again. To maximize the chances of
# successfully committing the retry, the client should execute the retry in the
# same session as the original attempt. The original session's lock priority
# increases with each consecutive abort, meaning that each attempt has a
# slightly better chance of success than the previous. Under some circumstances (
# e.g., many transactions attempting to modify the same row(s)), a transaction
# can abort many times in a short period before successfully committing. Thus,
# it is not a good idea to cap the number of retries a transaction can attempt;
# instead, it is better to limit the total amount of wall time spent retrying. ##
# # Idle Transactions A transaction is considered idle if it has no outstanding
# reads or SQL queries and has not started a read or SQL query within the last
# 10 seconds. Idle transactions can be aborted by Cloud Spanner so that they don'
# t hold on to locks indefinitely. In that case, the commit will fail with error
# `ABORTED`. If this behavior is undesirable, periodically executing a simple
# SQL query in the transaction (e.g., `SELECT 1`) prevents the transaction from
# becoming idle. ## Snapshot Read-Only Transactions Snapshot read-only
# transactions provides a simpler method than locking read-write transactions
# for doing several consistent reads. However, this type of transaction does not
# support writes. Snapshot transactions do not take locks. Instead, they work by
# choosing a Cloud Spanner timestamp, then executing all reads at that timestamp.
# Since they do not acquire locks, they do not block concurrent read-write
# transactions. Unlike locking read-write transactions, snapshot read-only
# transactions never abort. They can fail if the chosen read timestamp is
# garbage collected; however, the default garbage collection policy is generous
# enough that most applications do not need to worry about this in practice.
# Snapshot read-only transactions do not need to call Commit or Rollback (and in
# fact are not permitted to do so). To execute a snapshot transaction, the
# client specifies a timestamp bound, which tells Cloud Spanner how to choose a
# read timestamp. The types of timestamp bound are: - Strong (the default). -
# Bounded staleness. - Exact staleness. If the Cloud Spanner database to be read
# is geographically distributed, stale read-only transactions can execute more
# quickly than strong or read-write transaction, because they are able to
# execute far from the leader replica. Each type of timestamp bound is discussed
# in detail below. ### Strong Strong reads are guaranteed to see the effects of
# all transactions that have committed before the start of the read. Furthermore,
# all rows yielded by a single read are consistent with each other -- if any
# part of the read observes a transaction, all parts of the read see the
# transaction. Strong reads are not repeatable: two consecutive strong read-only
# transactions might return inconsistent results if there are concurrent writes.
# If consistency across reads is required, the reads should be executed within a
# transaction or at an exact read timestamp. See TransactionOptions.ReadOnly.
# strong. ### Exact Staleness These timestamp bounds execute reads at a user-
# specified timestamp. Reads at a timestamp are guaranteed to see a consistent
# prefix of the global transaction history: they observe modifications done by
# all transactions with a commit timestamp <= the read timestamp, and observe
# none of the modifications done by transactions with a larger commit timestamp.
# They will block until all conflicting transactions that may be assigned commit
# timestamps <= the read timestamp have finished. The timestamp can either be
# expressed as an absolute Cloud Spanner commit timestamp or a staleness
# relative to the current time. These modes do not require a "negotiation phase"
# to pick a timestamp. As a result, they execute slightly faster than the
# equivalent boundedly stale concurrency modes. On the other hand, boundedly
# stale reads usually return fresher results. See TransactionOptions.ReadOnly.
# read_timestamp and TransactionOptions.ReadOnly.exact_staleness. ### Bounded
# Staleness Bounded staleness modes allow Cloud Spanner to pick the read
# timestamp, subject to a user-provided staleness bound. Cloud Spanner chooses
# the newest timestamp within the staleness bound that allows execution of the
# reads at the closest available replica without blocking. All rows yielded are
# consistent with each other -- if any part of the read observes a transaction,
# all parts of the read see the transaction. Boundedly stale reads are not
# repeatable: two stale reads, even if they use the same staleness bound, can
# execute at different timestamps and thus return inconsistent results.
# Boundedly stale reads execute in two phases: the first phase negotiates a
# timestamp among all replicas needed to serve the read. In the second phase,
# reads are executed at the negotiated timestamp. As a result of the two phase
# execution, bounded staleness reads are usually a little slower than comparable
# exact staleness reads. However, they are typically able to return fresher
# results, and are more likely to execute at the closest replica. Because the
# timestamp negotiation requires up-front knowledge of which rows will be read,
# it can only be used with single-use read-only transactions. See
# TransactionOptions.ReadOnly.max_staleness and TransactionOptions.ReadOnly.
# min_read_timestamp. ### Old Read Timestamps and Garbage Collection Cloud
# Spanner continuously garbage collects deleted and overwritten data in the
# background to reclaim storage space. This process is known as "version GC". By
# default, version GC reclaims versions after they are one hour old. Because of
# this, Cloud Spanner cannot perform reads at read timestamps more than one hour
# in the past. This restriction also applies to in-progress reads and/or SQL
# queries whose timestamp become too old while executing. Reads and SQL queries
# with too-old read timestamps fail with the error `FAILED_PRECONDITION`. ##
# Partitioned DML Transactions Partitioned DML transactions are used to execute
# DML statements with a different execution strategy that provides different,
# and often better, scalability properties for large, table-wide operations than
# DML in a ReadWrite transaction. Smaller scoped statements, such as an OLTP
# workload, should prefer using ReadWrite transactions. Partitioned DML
# partitions the keyspace and runs the DML statement on each partition in
# separate, internal transactions. These transactions commit automatically when
# complete, and run independently from one another. To reduce lock contention,
# this execution strategy only acquires read locks on rows that match the WHERE
# clause of the statement. Additionally, the smaller per-partition transactions
# hold locks for less time. That said, Partitioned DML is not a drop-in
# replacement for standard DML used in ReadWrite transactions. - The DML
# statement must be fully-partitionable. Specifically, the statement must be
# expressible as the union of many statements which each access only a single
# row of the table. - The statement is not applied atomically to all rows of the
# table. Rather, the statement is applied atomically to partitions of the table,
# in independent transactions. Secondary index rows are updated atomically with
# the base table rows. - Partitioned DML does not guarantee exactly-once
# execution semantics against a partition. The statement will be applied at
# least once to each partition. It is strongly recommended that the DML
# statement should be idempotent to avoid unexpected results. For instance, it
# is potentially dangerous to run a statement such as `UPDATE table SET column =
# column + 1` as it could be run multiple times against some rows. - The
# partitions are committed automatically - there is no support for Commit or
# Rollback. If the call returns an error, or if the client issuing the
# ExecuteSql call dies, it is possible that some rows had the statement executed
# on them successfully. It is also possible that statement was never executed
# against other rows. - Partitioned DML transactions may only contain the
# execution of a single DML statement via ExecuteSql or ExecuteStreamingSql. -
# If any error is encountered during the execution of the partitioned DML
# operation (for instance, a UNIQUE INDEX violation, division by zero, or a
# value that cannot be stored due to schema constraints), then the operation is
# stopped at that point and an error is returned. It is possible that at this
# point, some partitions have been committed (or even committed multiple times),
# and other partitions have not been run at all. Given the above, Partitioned
# DML is good fit for large, database-wide, operations that are idempotent, such
# as deleting old rows from a very large table.
# Corresponds to the JSON property `singleUse`
# @return [Google::Apis::SpannerV1::TransactionOptions]
attr_accessor :single_use
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@begin = args[:begin] if args.key?(:begin)
@id = args[:id] if args.key?(:id)
@single_use = args[:single_use] if args.key?(:single_use)
end
end
# `Type` indicates the type of a Cloud Spanner value, as might be stored in a
# table cell or returned from an SQL query.
class Type
include Google::Apis::Core::Hashable
# `Type` indicates the type of a Cloud Spanner value, as might be stored in a
# table cell or returned from an SQL query.
# Corresponds to the JSON property `arrayElementType`
# @return [Google::Apis::SpannerV1::Type]
attr_accessor :array_element_type
# Required. The TypeCode for this type.
# Corresponds to the JSON property `code`
# @return [String]
attr_accessor :code
# `StructType` defines the fields of a STRUCT type.
# Corresponds to the JSON property `structType`
# @return [Google::Apis::SpannerV1::StructType]
attr_accessor :struct_type
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@array_element_type = args[:array_element_type] if args.key?(:array_element_type)
@code = args[:code] if args.key?(:code)
@struct_type = args[:struct_type] if args.key?(:struct_type)
end
end
# Metadata type for the operation returned by UpdateDatabaseDdl.
class UpdateDatabaseDdlMetadata
include Google::Apis::Core::Hashable
# Reports the commit timestamps of all statements that have succeeded so far,
# where `commit_timestamps[i]` is the commit timestamp for the statement `
# statements[i]`.
# Corresponds to the JSON property `commitTimestamps`
# @return [Array<String>]
attr_accessor :commit_timestamps
# The database being modified.
# Corresponds to the JSON property `database`
# @return [String]
attr_accessor :database
# For an update this list contains all the statements. For an individual
# statement, this list contains only that statement.
# Corresponds to the JSON property `statements`
# @return [Array<String>]
attr_accessor :statements
# Output only. When true, indicates that the operation is throttled e.g due to
# resource constraints. When resources become available the operation will
# resume and this field will be false again.
# Corresponds to the JSON property `throttled`
# @return [Boolean]
attr_accessor :throttled
alias_method :throttled?, :throttled
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@commit_timestamps = args[:commit_timestamps] if args.key?(:commit_timestamps)
@database = args[:database] if args.key?(:database)
@statements = args[:statements] if args.key?(:statements)
@throttled = args[:throttled] if args.key?(:throttled)
end
end
# Enqueues the given DDL statements to be applied, in order but not necessarily
# all at once, to the database schema at some point (or points) in the future.
# The server checks that the statements are executable (syntactically valid,
# name tables that exist, etc.) before enqueueing them, but they may still fail
# upon later execution (e.g., if a statement from another batch of statements is
# applied first and it conflicts in some way, or if there is some data-related
# problem like a `NULL` value in a column to which `NOT NULL` would be added).
# If a statement fails, all subsequent statements in the batch are automatically
# cancelled. Each batch of statements is assigned a name which can be used with
# the Operations API to monitor progress. See the operation_id field for more
# details.
class UpdateDatabaseDdlRequest
include Google::Apis::Core::Hashable
# If empty, the new update request is assigned an automatically-generated
# operation ID. Otherwise, `operation_id` is used to construct the name of the
# resulting Operation. Specifying an explicit operation ID simplifies
# determining whether the statements were executed in the event that the
# UpdateDatabaseDdl call is replayed, or the return value is otherwise lost: the
# database and `operation_id` fields can be combined to form the name of the
# resulting longrunning.Operation: `/operations/`. `operation_id` should be
# unique within the database, and must be a valid identifier: `a-z*`. Note that
# automatically-generated operation IDs always begin with an underscore. If the
# named operation already exists, UpdateDatabaseDdl returns `ALREADY_EXISTS`.
# Corresponds to the JSON property `operationId`
# @return [String]
attr_accessor :operation_id
# Required. DDL statements to be applied to the database.
# Corresponds to the JSON property `statements`
# @return [Array<String>]
attr_accessor :statements
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@operation_id = args[:operation_id] if args.key?(:operation_id)
@statements = args[:statements] if args.key?(:statements)
end
end
# Metadata type for the operation returned by UpdateInstance.
class UpdateInstanceMetadata
include Google::Apis::Core::Hashable
# The time at which this operation was cancelled. If set, this operation is in
# the process of undoing itself (which is guaranteed to succeed) and cannot be
# cancelled again.
# Corresponds to the JSON property `cancelTime`
# @return [String]
attr_accessor :cancel_time
# The time at which this operation failed or was completed successfully.
# Corresponds to the JSON property `endTime`
# @return [String]
attr_accessor :end_time
# An isolated set of Cloud Spanner resources on which databases can be hosted.
# Corresponds to the JSON property `instance`
# @return [Google::Apis::SpannerV1::Instance]
attr_accessor :instance
# The time at which UpdateInstance request was received.
# Corresponds to the JSON property `startTime`
# @return [String]
attr_accessor :start_time
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@cancel_time = args[:cancel_time] if args.key?(:cancel_time)
@end_time = args[:end_time] if args.key?(:end_time)
@instance = args[:instance] if args.key?(:instance)
@start_time = args[:start_time] if args.key?(:start_time)
end
end
# The request for UpdateInstance.
class UpdateInstanceRequest
include Google::Apis::Core::Hashable
# Required. A mask specifying which fields in Instance should be updated. The
# field mask must always be specified; this prevents any future fields in
# Instance from being erased accidentally by clients that do not know about them.
# Corresponds to the JSON property `fieldMask`
# @return [String]
attr_accessor :field_mask
# An isolated set of Cloud Spanner resources on which databases can be hosted.
# Corresponds to the JSON property `instance`
# @return [Google::Apis::SpannerV1::Instance]
attr_accessor :instance
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@field_mask = args[:field_mask] if args.key?(:field_mask)
@instance = args[:instance] if args.key?(:instance)
end
end
# Arguments to insert, update, insert_or_update, and replace operations.
class Write
include Google::Apis::Core::Hashable
# The names of the columns in table to be written. The list of columns must
# contain enough columns to allow Cloud Spanner to derive values for all primary
# key columns in the row(s) to be modified.
# Corresponds to the JSON property `columns`
# @return [Array<String>]
attr_accessor :columns
# Required. The table whose rows will be written.
# Corresponds to the JSON property `table`
# @return [String]
attr_accessor :table
# The values to be written. `values` can contain more than one list of values.
# If it does, then multiple rows are written, one for each entry in `values`.
# Each list in `values` must have exactly as many entries as there are entries
# in columns above. Sending multiple lists is equivalent to sending multiple `
# Mutation`s, each containing one `values` entry and repeating table and columns.
# Individual values in each list are encoded as described here.
# Corresponds to the JSON property `values`
# @return [Array<Array<Object>>]
attr_accessor :values
def initialize(**args)
update!(**args)
end
# Update properties of this object
def update!(**args)
@columns = args[:columns] if args.key?(:columns)
@table = args[:table] if args.key?(:table)
@values = args[:values] if args.key?(:values)
end
end
end
end
end