| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import enum |
| |
|
| |
|
| | class AutoRowIDs(enum.Enum): |
| | """How to handle automatic insert IDs when inserting rows as a stream.""" |
| |
|
| | DISABLED = enum.auto() |
| | GENERATE_UUID = enum.auto() |
| |
|
| |
|
| | class Compression(str, enum.Enum): |
| | """The compression type to use for exported files. The default value is |
| | :attr:`NONE`. |
| | |
| | :attr:`DEFLATE` and :attr:`SNAPPY` are |
| | only supported for Avro. |
| | """ |
| |
|
| | GZIP = "GZIP" |
| | """Specifies GZIP format.""" |
| |
|
| | DEFLATE = "DEFLATE" |
| | """Specifies DEFLATE format.""" |
| |
|
| | SNAPPY = "SNAPPY" |
| | """Specifies SNAPPY format.""" |
| |
|
| | ZSTD = "ZSTD" |
| | """Specifies ZSTD format.""" |
| |
|
| | NONE = "NONE" |
| | """Specifies no compression.""" |
| |
|
| |
|
| | class DecimalTargetType: |
| | """The data types that could be used as a target type when converting decimal values. |
| | |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#DecimalTargetType |
| | |
| | .. versionadded:: 2.21.0 |
| | """ |
| |
|
| | NUMERIC = "NUMERIC" |
| | """Decimal values could be converted to NUMERIC type.""" |
| |
|
| | BIGNUMERIC = "BIGNUMERIC" |
| | """Decimal values could be converted to BIGNUMERIC type.""" |
| |
|
| | STRING = "STRING" |
| | """Decimal values could be converted to STRING type.""" |
| |
|
| |
|
| | class CreateDisposition(object): |
| | """Specifies whether the job is allowed to create new tables. The default |
| | value is :attr:`CREATE_IF_NEEDED`. |
| | |
| | Creation, truncation and append actions occur as one atomic update |
| | upon job completion. |
| | """ |
| |
|
| | CREATE_IF_NEEDED = "CREATE_IF_NEEDED" |
| | """If the table does not exist, BigQuery creates the table.""" |
| |
|
| | CREATE_NEVER = "CREATE_NEVER" |
| | """The table must already exist. If it does not, a 'notFound' error is |
| | returned in the job result.""" |
| |
|
| |
|
| | class DefaultPandasDTypes(enum.Enum): |
| | """Default Pandas DataFrem DTypes to convert BigQuery data. These |
| | Sentinel values are used instead of None to maintain backward compatibility, |
| | and allow Pandas package is not available. For more information: |
| | https://stackoverflow.com/a/60605919/101923 |
| | """ |
| |
|
| | BOOL_DTYPE = object() |
| | """Specifies default bool dtype""" |
| |
|
| | INT_DTYPE = object() |
| | """Specifies default integer dtype""" |
| |
|
| | DATE_DTYPE = object() |
| | """Specifies default date dtype""" |
| |
|
| | TIME_DTYPE = object() |
| | """Specifies default time dtype""" |
| |
|
| | RANGE_DATE_DTYPE = object() |
| | """Specifies default range date dtype""" |
| |
|
| | RANGE_DATETIME_DTYPE = object() |
| | """Specifies default range datetime dtype""" |
| |
|
| | RANGE_TIMESTAMP_DTYPE = object() |
| | """Specifies default range timestamp dtype""" |
| |
|
| |
|
| | class DestinationFormat(object): |
| | """The exported file format. The default value is :attr:`CSV`. |
| | |
| | Tables with nested or repeated fields cannot be exported as CSV. |
| | """ |
| |
|
| | CSV = "CSV" |
| | """Specifies CSV format.""" |
| |
|
| | NEWLINE_DELIMITED_JSON = "NEWLINE_DELIMITED_JSON" |
| | """Specifies newline delimited JSON format.""" |
| |
|
| | AVRO = "AVRO" |
| | """Specifies Avro format.""" |
| |
|
| | PARQUET = "PARQUET" |
| | """Specifies Parquet format.""" |
| |
|
| |
|
| | class Encoding(object): |
| | """The character encoding of the data. The default is :attr:`UTF_8`. |
| | |
| | BigQuery decodes the data after the raw, binary data has been |
| | split using the values of the quote and fieldDelimiter properties. |
| | """ |
| |
|
| | UTF_8 = "UTF-8" |
| | """Specifies UTF-8 encoding.""" |
| |
|
| | ISO_8859_1 = "ISO-8859-1" |
| | """Specifies ISO-8859-1 encoding.""" |
| |
|
| |
|
| | class QueryPriority(object): |
| | """Specifies a priority for the query. The default value is |
| | :attr:`INTERACTIVE`. |
| | """ |
| |
|
| | INTERACTIVE = "INTERACTIVE" |
| | """Specifies interactive priority.""" |
| |
|
| | BATCH = "BATCH" |
| | """Specifies batch priority.""" |
| |
|
| |
|
| | class QueryApiMethod(str, enum.Enum): |
| | """API method used to start the query. The default value is |
| | :attr:`INSERT`. |
| | """ |
| |
|
| | INSERT = "INSERT" |
| | """Submit a query job by using the `jobs.insert REST API method |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/insert>`_. |
| | |
| | This supports all job configuration options. |
| | """ |
| |
|
| | QUERY = "QUERY" |
| | """Submit a query job by using the `jobs.query REST API method |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/query>`_. |
| | |
| | Differences from ``INSERT``: |
| | |
| | * Many parameters and job configuration options, including job ID and |
| | destination table, cannot be used |
| | with this API method. See the `jobs.query REST API documentation |
| | <https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/query>`_ for |
| | the complete list of supported configuration options. |
| | |
| | * API blocks up to a specified timeout, waiting for the query to |
| | finish. |
| | |
| | * The full job resource (including job statistics) may not be available. |
| | Call :meth:`~google.cloud.bigquery.job.QueryJob.reload` or |
| | :meth:`~google.cloud.bigquery.client.Client.get_job` to get full job |
| | statistics and configuration. |
| | |
| | * :meth:`~google.cloud.bigquery.Client.query` can raise API exceptions if |
| | the query fails, whereas the same errors don't appear until calling |
| | :meth:`~google.cloud.bigquery.job.QueryJob.result` when the ``INSERT`` |
| | API method is used. |
| | """ |
| |
|
| |
|
| | class SchemaUpdateOption(object): |
| | """Specifies an update to the destination table schema as a side effect of |
| | a load job. |
| | """ |
| |
|
| | ALLOW_FIELD_ADDITION = "ALLOW_FIELD_ADDITION" |
| | """Allow adding a nullable field to the schema.""" |
| |
|
| | ALLOW_FIELD_RELAXATION = "ALLOW_FIELD_RELAXATION" |
| | """Allow relaxing a required field in the original schema to nullable.""" |
| |
|
| |
|
| | class SourceFormat(object): |
| | """The format of the data files. The default value is :attr:`CSV`. |
| | |
| | Note that the set of allowed values for loading data is different |
| | than the set used for external data sources (see |
| | :class:`~google.cloud.bigquery.external_config.ExternalSourceFormat`). |
| | """ |
| |
|
| | CSV = "CSV" |
| | """Specifies CSV format.""" |
| |
|
| | DATASTORE_BACKUP = "DATASTORE_BACKUP" |
| | """Specifies datastore backup format""" |
| |
|
| | NEWLINE_DELIMITED_JSON = "NEWLINE_DELIMITED_JSON" |
| | """Specifies newline delimited JSON format.""" |
| |
|
| | AVRO = "AVRO" |
| | """Specifies Avro format.""" |
| |
|
| | PARQUET = "PARQUET" |
| | """Specifies Parquet format.""" |
| |
|
| | ORC = "ORC" |
| | """Specifies Orc format.""" |
| |
|
| |
|
| | class KeyResultStatementKind: |
| | """Determines which statement in the script represents the "key result". |
| | |
| | The "key result" is used to populate the schema and query results of the script job. |
| | |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#keyresultstatementkind |
| | """ |
| |
|
| | KEY_RESULT_STATEMENT_KIND_UNSPECIFIED = "KEY_RESULT_STATEMENT_KIND_UNSPECIFIED" |
| | LAST = "LAST" |
| | FIRST_SELECT = "FIRST_SELECT" |
| |
|
| |
|
| | class StandardSqlTypeNames(str, enum.Enum): |
| | def _generate_next_value_(name, start, count, last_values): |
| | return name |
| |
|
| | TYPE_KIND_UNSPECIFIED = enum.auto() |
| | INT64 = enum.auto() |
| | BOOL = enum.auto() |
| | FLOAT64 = enum.auto() |
| | STRING = enum.auto() |
| | BYTES = enum.auto() |
| | TIMESTAMP = enum.auto() |
| | DATE = enum.auto() |
| | TIME = enum.auto() |
| | DATETIME = enum.auto() |
| | INTERVAL = enum.auto() |
| | GEOGRAPHY = enum.auto() |
| | NUMERIC = enum.auto() |
| | BIGNUMERIC = enum.auto() |
| | JSON = enum.auto() |
| | ARRAY = enum.auto() |
| | STRUCT = enum.auto() |
| | RANGE = enum.auto() |
| |
|
| |
|
| | class EntityTypes(str, enum.Enum): |
| | """Enum of allowed entity type names in AccessEntry""" |
| |
|
| | USER_BY_EMAIL = "userByEmail" |
| | GROUP_BY_EMAIL = "groupByEmail" |
| | DOMAIN = "domain" |
| | DATASET = "dataset" |
| | SPECIAL_GROUP = "specialGroup" |
| | VIEW = "view" |
| | IAM_MEMBER = "iamMember" |
| | ROUTINE = "routine" |
| |
|
| |
|
| | |
| | |
| | class SqlTypeNames(str, enum.Enum): |
| | """Enum of allowed SQL type names in schema.SchemaField.""" |
| |
|
| | STRING = "STRING" |
| | BYTES = "BYTES" |
| | INTEGER = "INTEGER" |
| | INT64 = "INTEGER" |
| | FLOAT = "FLOAT" |
| | FLOAT64 = "FLOAT" |
| | DECIMAL = NUMERIC = "NUMERIC" |
| | BIGDECIMAL = BIGNUMERIC = "BIGNUMERIC" |
| | BOOLEAN = "BOOLEAN" |
| | BOOL = "BOOLEAN" |
| | GEOGRAPHY = "GEOGRAPHY" |
| | RECORD = "RECORD" |
| | STRUCT = "RECORD" |
| | TIMESTAMP = "TIMESTAMP" |
| | DATE = "DATE" |
| | TIME = "TIME" |
| | DATETIME = "DATETIME" |
| | INTERVAL = "INTERVAL" |
| | RANGE = "RANGE" |
| |
|
| |
|
| | class WriteDisposition(object): |
| | """Specifies the action that occurs if destination table already exists. |
| | |
| | The default value is :attr:`WRITE_APPEND`. |
| | |
| | Each action is atomic and only occurs if BigQuery is able to complete |
| | the job successfully. Creation, truncation and append actions occur as one |
| | atomic update upon job completion. |
| | """ |
| |
|
| | WRITE_APPEND = "WRITE_APPEND" |
| | """If the table already exists, BigQuery appends the data to the table.""" |
| |
|
| | WRITE_TRUNCATE = "WRITE_TRUNCATE" |
| | """If the table already exists, BigQuery overwrites the table data.""" |
| |
|
| | WRITE_EMPTY = "WRITE_EMPTY" |
| | """If the table already exists and contains data, a 'duplicate' error is |
| | returned in the job result.""" |
| |
|
| |
|
| | class DeterminismLevel: |
| | """Specifies determinism level for JavaScript user-defined functions (UDFs). |
| | |
| | https://cloud.google.com/bigquery/docs/reference/rest/v2/routines#DeterminismLevel |
| | """ |
| |
|
| | DETERMINISM_LEVEL_UNSPECIFIED = "DETERMINISM_LEVEL_UNSPECIFIED" |
| | """The determinism of the UDF is unspecified.""" |
| |
|
| | DETERMINISTIC = "DETERMINISTIC" |
| | """The UDF is deterministic, meaning that 2 function calls with the same inputs |
| | always produce the same result, even across 2 query runs.""" |
| |
|
| | NOT_DETERMINISTIC = "NOT_DETERMINISTIC" |
| | """The UDF is not deterministic.""" |
| |
|