Class Client (3.39.0)

Client(project:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,_http:typing.Optional[requests.sessions.Session]=None,location:typing.Optional[str]=None,default_query_job_config:typing.Optional[google.cloud.bigquery.job.query.QueryJobConfig]=None,default_load_job_config:typing.Optional[google.cloud.bigquery.job.load.LoadJobConfig]=None,client_info:typing.Optional[google.api_core.client_info.ClientInfo]=None,client_options:typing.Optional[typing.Union[google.api_core.client_options.ClientOptions,typing.Dict[str,typing.Any]]]=None,default_job_creation_mode:typing.Optional[str]=None,)

Client to bundle configuration needed for API requests.

Parameters

NameDescription
projectOptional[str]

Project ID for the project which the client acts on behalf of. Will be passed when creating a dataset / job. If not passed, falls back to the default inferred from the environment.

credentialsOptional[google.auth.credentials.Credentials]

The OAuth2 Credentials to use for this client. If not passed (and if no_http object is passed), falls back to the default inferred from the environment.

_httpOptional[requests.Session]

HTTP object to make requests. Can be any object that definesrequest() with the same interface asrequests.Session.request. If not passed, an_http object is created that is bound to thecredentials for the current object. This parameter should be considered private, and could change in the future.

locationOptional[str]

Default location for jobs / datasets / tables.

default_query_job_configOptional[google.cloud.bigquery.job.QueryJobConfig]

DefaultQueryJobConfig. Will be merged into job configs passed into thequery method.

default_load_job_configOptional[google.cloud.bigquery.job.LoadJobConfig]

DefaultLoadJobConfig. Will be merged into job configs passed into theload_table_* methods.

client_infoOptional[google.api_core.client_info.ClientInfo]

The client info used to send a user-agent string along with API requests. IfNone, then default info will be used. Generally, you only need to set this if you're developing your own library or partner tool.

client_optionsOptional[Union[google.api_core.client_options.ClientOptions, Dict]]

Client options used to set user options on the client. API Endpoint should be set through client_options.

default_job_creation_modeOptional[str]

Sets the default job creation mode used by query methods such as query_and_wait(). For lightweight queries, JOB_CREATION_OPTIONAL is generally recommended.

Properties

default_job_creation_mode

Default job creation mode used for query execution.

default_load_job_config

DefaultLoadJobConfig.Will be merged into job configs passed into theload_table_* methods.

default_query_job_config

DefaultQueryJobConfig orNone.

Will be merged into job configs passed into thequery orquery_and_wait methods.

location

Default location for jobs / datasets / tables.

Methods

cancel_job

cancel_job(job_id:str,project:typing.Optional[str]=None,location:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->typing.Union[google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.query.QueryJob,]
Parameters
NameDescription
job_idUnion[ str,google.cloud.bigquery.job.LoadJob,google.cloud.bigquery.job.CopyJob,google.cloud.bigquery.job.ExtractJob,google.cloud.bigquery.job.QueryJob ]

Job identifier.

projectOptional[str]

ID of the project which owns the job (defaults to the client's project).

locationOptional[str]

Location where the job was run. Ignored ifjob_id is a job object.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
Union[google.cloud.bigquery.job.LoadJob,google.cloud.bigquery.job.CopyJob,google.cloud.bigquery.job.ExtractJob,google.cloud.bigquery.job.QueryJob, ]Job instance, based on the resource returned by the API.

close

close()

copy_table

copy_table(sources:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,typing.Sequence[typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,]],],destination:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,job_config:typing.Optional[google.cloud.bigquery.job.copy_.CopyJobConfig]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.job.copy_.CopyJob
Parameters
NameDescription
sourcesUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, Sequence[ Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ] ], ]

Table or tables to be copied.

destinationUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

Table into which data is to be copied.

job_idOptional[str]

The ID of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of any source table as well as the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[google.cloud.bigquery.job.CopyJobConfig]

Extra configuration options for the job.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofCopyJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.job.CopyJobA new copy job instance.

create_dataset

create_dataset(dataset:typing.Union[str,google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem,],exists_ok:bool=False,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.dataset.Dataset

API call: create the dataset via a POST request.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/datasets/insert

Example:

from google.cloud importbigqueryclient =bigquery.Client()dataset =bigquery.Dataset('my_project.my_dataset')dataset = client.create_dataset(dataset)
Parameters
NameDescription
datasetUnion[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem, str, ]

ADataset to create. Ifdataset is a reference, an empty dataset is created with the specified ID and client's default location.

exists_okOptional[bool]

Defaults toFalse. IfTrue, ignore "already exists" errors when creating the dataset.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
google.cloud.exceptions.ConflictIf the dataset already exists.
Returns
TypeDescription
google.cloud.bigquery.dataset.DatasetA newDataset returned from the API.

create_job

create_job(job_config:dict,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->typing.Union[google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.query.QueryJob,]

Create a new job.

Parameters
NameDescription
job_configdict

configuration job representation returned from the API.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
Union[google.cloud.bigquery.job.LoadJob,google.cloud.bigquery.job.CopyJob,google.cloud.bigquery.job.ExtractJob,google.cloud.bigquery.job.QueryJob ]A new job instance.

create_routine

create_routine(routine:google.cloud.bigquery.routine.routine.Routine,exists_ok:bool=False,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.routine.routine.Routine
Parameters
NameDescription
routinegoogle.cloud.bigquery.routine.Routine

ARoutine to create. The dataset that the routine belongs to must already exist.

exists_okOptional[bool]

Defaults toFalse. IfTrue, ignore "already exists" errors when creating the routine.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
google.cloud.exceptions.ConflictIf the routine already exists.
Returns
TypeDescription
google.cloud.bigquery.routine.RoutineA newRoutine returned from the service.

create_table

create_table(table:typing.Union[str,google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,],exists_ok:bool=False,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.table.Table
Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

ATable to create. Iftable is a reference, an empty table is created with the specified ID. The dataset that the table belongs to must already exist.

exists_okOptional[bool]

Defaults toFalse. IfTrue, ignore "already exists" errors when creating the table.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
google.cloud.exceptions.ConflictIf the table already exists.
Returns
TypeDescription
google.cloud.bigquery.table.TableA newTable returned from the service.

dataset

dataset(dataset_id:str,project:typing.Optional[str]=None)->google.cloud.bigquery.dataset.DatasetReference
Parameters
NameDescription
dataset_idstr

ID of the dataset.

projectOptional[str]

Project ID for the dataset (defaults to the project of the client).

Returns
TypeDescription
google.cloud.bigquery.dataset.DatasetReferencea newDatasetReference instance.

delete_dataset

Parameters
NameDescription
datasetUnion[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem, str, ]

A reference to the dataset to delete. If a string is passed in, this method attempts to create a dataset reference from a string usingfrom_string.

delete_contentsOptional[bool]

If True, delete all the tables in the dataset. If False and the dataset contains tables, the request will fail. Default is False.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

not_found_okOptional[bool]

Defaults toFalse. IfTrue, ignore "not found" errors when deleting the dataset.

delete_job_metadata

delete_job_metadata(job_id:typing.Union[str,google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.query.QueryJob,],project:typing.Optional[str]=None,location:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,not_found_ok:bool=False,)

[Beta] Delete job metadata from job history.

Note: This does not stop a running job. Usexref_cancel_job instead.

Parameters
NameDescription
job_idUnion[ str, LoadJob, CopyJob, ExtractJob, QueryJob ]

Job or job identifier.

projectOptional[str]

ID of the project which owns the job (defaults to the client's project).

locationOptional[str]

Location where the job was run. Ignored ifjob_id is a job object.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

not_found_okOptional[bool]

Defaults toFalse. IfTrue, ignore "not found" errors when deleting the job.

delete_model

delete_model(model:typing.Union[google.cloud.bigquery.model.Model,google.cloud.bigquery.model.ModelReference,str,],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,not_found_ok:bool=False,)->None
Parameters
NameDescription
modelUnion[google.cloud.bigquery.model.Model,google.cloud.bigquery.model.ModelReference, str, ]

A reference to the model to delete. If a string is passed in, this method attempts to create a model reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

not_found_okOptional[bool]

Defaults toFalse. IfTrue, ignore "not found" errors when deleting the model.

delete_routine

delete_routine(routine:typing.Union[google.cloud.bigquery.routine.routine.Routine,google.cloud.bigquery.routine.routine.RoutineReference,str,],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,not_found_ok:bool=False,)->None
Parameters
NameDescription
routineUnion[google.cloud.bigquery.routine.Routine,google.cloud.bigquery.routine.RoutineReference, str, ]

A reference to the routine to delete. If a string is passed in, this method attempts to create a routine reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

not_found_okOptional[bool]

Defaults toFalse. IfTrue, ignore "not found" errors when deleting the routine.

delete_table

delete_table(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,not_found_ok:bool=False,)->None
Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

A reference to the table to delete. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

not_found_okOptional[bool]

Defaults toFalse. IfTrue, ignore "not found" errors when deleting the table.

extract_table

extract_table(source:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,google.cloud.bigquery.model.Model,google.cloud.bigquery.model.ModelReference,str,],destination_uris:typing.Union[str,typing.Sequence[str]],job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,job_config:typing.Optional[google.cloud.bigquery.job.extract.ExtractJobConfig]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,source_type:str="Table",)->google.cloud.bigquery.job.extract.ExtractJob

Start a job to extract a table into Cloud Storage files.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/Job#jobconfigurationextract

Parameters
NameDescription
sourceUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,google.cloud.bigquery.model.Model,google.cloud.bigquery.model.ModelReference, src, ]

Table or Model to be extracted.

destination_urisUnion[str, Sequence[str]]

URIs of Cloud Storage file(s) into which table data is to be extracted; in formatgs://<bucket_name>/<object_name_or_glob>.

job_idOptional[str]

The ID of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the source table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[google.cloud.bigquery.job.ExtractJobConfig]

Extra configuration options for the job.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

source_typeOptional[str]

Type of source to be extracted.Table orModel. Defaults toTable.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofExtractJobConfig class.
ValueErrorIfsource_type is not amongTable,Model.
Returns
TypeDescription
google.cloud.bigquery.job.ExtractJobA new extract job instance.

get_dataset

get_dataset(dataset_ref:typing.Union[google.cloud.bigquery.dataset.DatasetReference,str],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,dataset_view:typing.Optional[google.cloud.bigquery.enums.DatasetView]=None,)->google.cloud.bigquery.dataset.Dataset

Fetch the dataset referenced bydataset_ref

Parameters
NameDescription
dataset_refUnion[google.cloud.bigquery.dataset.DatasetReference, str, ]

A reference to the dataset to fetch from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

dataset_viewOptional[google.cloud.bigquery.enums.DatasetView]

Specifies the view that determines which dataset information is returned. By default, dataset metadata (e.g. friendlyName, description, labels, etc) and ACL information are returned. This argument can take on the following possible enum values. *ACL: Includes dataset metadata and the ACL. *FULL: Includes all dataset metadata, including the ACL and table metadata. This view is not supported by thedatasets.list API method. *METADATA: Includes basic dataset metadata, but not the ACL. *DATASET_VIEW_UNSPECIFIED: The server will decide which view to use. Currently defaults to FULL.

Returns
TypeDescription
google.cloud.bigquery.dataset.DatasetADataset instance.

get_iam_policy

get_iam_policy(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],requested_policy_version:int=1,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.api_core.iam.Policy

Return the access control policy for a table resource.

Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

The table to get the access control policy for. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

requested_policy_versionint

Optional. The maximum policy version that will be used to format the policy. Only version1 is currently supported. See:https://cloud.google.com/bigquery/docs/reference/rest/v2/GetPolicyOptions

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.api_core.iam.PolicyThe access control policy.

get_job

get_job(job_id:typing.Union[str,google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.query.QueryJob,],project:typing.Optional[str]=None,location:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=128,)->typing.Union[google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.query.QueryJob,google.cloud.bigquery.job.base.UnknownJob,]

Fetch a job for the project associated with this client.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/get

Parameters
NameDescription
job_idUnion[ str, job.LoadJob, job.CopyJob, job.ExtractJob, job.QueryJob ]

Job identifier.

projectOptional[str]

ID of the project which owns the job (defaults to the client's project).

locationOptional[str]

Location where the job was run. Ignored ifjob_id is a job object.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
Union[job.LoadJob, job.CopyJob, job.ExtractJob, job.QueryJob, job.UnknownJob]Job instance, based on the resource returned by the API.

get_model

get_model(model_ref:typing.Union[google.cloud.bigquery.model.ModelReference,str],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.model.Model

[Beta] Fetch the model referenced bymodel_ref.

Parameters
NameDescription
model_refUnion[google.cloud.bigquery.model.ModelReference, str, ]

A reference to the model to fetch from the BigQuery API. If a string is passed in, this method attempts to create a model reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.model.ModelAModel instance.

get_routine

get_routine(routine_ref:typing.Union[google.cloud.bigquery.routine.routine.Routine,google.cloud.bigquery.routine.routine.RoutineReference,str,],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.routine.routine.Routine

[Beta] Get the routine referenced byroutine_ref.

Parameters
NameDescription
routine_refUnion[google.cloud.bigquery.routine.Routine,google.cloud.bigquery.routine.RoutineReference, str, ]

A reference to the routine to fetch from the BigQuery API. If a string is passed in, this method attempts to create a reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the API call.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.routine.RoutineARoutine instance.

get_service_account_email

get_service_account_email(project:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->str

Get the email address of the project's BigQuery service account

Example:

from google.cloud importbigqueryclient =bigquery.Client()client.get_service_account_email()# returns an email similar to: my_service_account@my-project.iam.gserviceaccount.com
Note:This is the service account that BigQuery uses to manage tablesencrypted by a key in KMS.
Parameters
NameDescription
projectOptional[str]

Project ID to use for retreiving service account email. Defaults to the client's project.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
strservice account email address

get_table

Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

A reference to the table to fetch from the BigQuery API. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.table.TableATable instance.

insert_rows

Insert rows into a table via the streaming API.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/tabledata/insertAll

BigQuery will reject insertAll payloads that exceed a defined limit (10MB).Additionally, if a payload vastly exceeds this limit, the request is rejectedby the intermediate architecture, which returns a 413 (Payload Too Large) status code.

Seehttps://cloud.google.com/bigquery/quotas#streaming_inserts

Parameters
NameDescription
kwargsdict

Keyword arguments toinsert_rows_json.

tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference, str, ]

The destination table for the row data, or a reference to it.

rowsUnion[Sequence[Tuple], Sequence[Dict]]

Row data to be inserted. If a list of tuples is given, each tuple should contain data for each schema field on the current table and in the same order as the schema fields. If a list of dictionaries is given, the keys must include all required fields in the schema. Keys which do not correspond to a field in the schema are ignored.

selected_fieldsSequence[google.cloud.bigquery.schema.SchemaField]

The fields to return. Required iftable is aTableReference.

Exceptions
TypeDescription
ValueErrorif table's schema is not set orrows is not aSequence.
Returns
TypeDescription
Sequence[Mappings]One mapping per row with insert errors: the "index" key identifies the row, and the "errors" key contains a list of the mappings describing one or more problems with the row.

insert_rows_from_dataframe

insert_rows_from_dataframe(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,str,],dataframe,selected_fields:typing.Optional[typing.Sequence[google.cloud.bigquery.schema.SchemaField]]=None,chunk_size:int=500,**kwargs:typing.Dict)->typing.Sequence[typing.Sequence[dict]]

Insert rows into a table from a dataframe via the streaming API.

BigQuery will reject insertAll payloads that exceed a defined limit (10MB).Additionally, if a payload vastly exceeds this limit, the request is rejectedby the intermediate architecture, which returns a 413 (Payload Too Large) status code.

Seehttps://cloud.google.com/bigquery/quotas#streaming_inserts

Parameters
NameDescription
kwargsDict

Keyword arguments toinsert_rows_json.

tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference, str, ]

The destination table for the row data, or a reference to it.

dataframepandas.DataFrame

Apandas.DataFrame containing the data to load. AnyNaN values present in the dataframe are omitted from the streaming API request(s).

selected_fieldsSequence[google.cloud.bigquery.schema.SchemaField]

The fields to return. Required iftable is aTableReference.

chunk_sizeint

The number of rows to stream in a single chunk. Must be positive.

Exceptions
TypeDescription
ValueErrorif table's schema is not set
Returns
TypeDescription
Sequence[Sequence[Mappings]]A list with insert errors for each insert chunk. Each element is a list containing one mapping per row with insert errors: the "index" key identifies the row, and the "errors" key contains a list of the mappings describing one or more problems with the row.

insert_rows_json

insert_rows_json(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],json_rows:typing.Sequence[typing.Mapping[str,typing.Any]],row_ids:typing.Optional[typing.Union[typing.Iterable[typing.Optional[str]],google.cloud.bigquery.enums.AutoRowIDs,]]=AutoRowIDs.GENERATE_UUID,skip_invalid_rows:typing.Optional[bool]=None,ignore_unknown_values:typing.Optional[bool]=None,template_suffix:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->typing.Sequence[dict]

Insert rows into a table without applying local type conversions.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/tabledata/insertAll

BigQuery will reject insertAll payloads that exceed a defined limit (10MB).Additionally, if a payload vastly exceeds this limit, the request is rejectedby the intermediate architecture, which returns a 413 (Payload Too Large) status code.

Seehttps://cloud.google.com/bigquery/quotas#streaming_inserts

Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Tablegoogle.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str ]

The destination table for the row data, or a reference to it.

json_rowsSequence[Dict]

Row data to be inserted. Keys must match the table schema fields and values must be JSON-compatible representations.

row_idsUnion[Iterable[str], AutoRowIDs, None]

Unique IDs, one per row being inserted. An ID can also beNone, indicating that an explicit insert ID shouldnot be used for that row. If the argument is omitted altogether, unique IDs are created automatically. .. versionchanged:: 2.21.0 Can also be an iterable, not just a sequence, or anAutoRowIDs enum member. .. deprecated:: 2.21.0 PassingNone to explicitly request autogenerating insert IDs is deprecated, useAutoRowIDs.GENERATE_UUID instead.

skip_invalid_rowsOptional[bool]

Insert all valid rows of a request, even if invalid rows exist. The default value isFalse, which causes the entire request to fail if any invalid rows exist.

ignore_unknown_valuesOptional[bool]

Accept rows that contain values that do not match the schema. The unknown values are ignored. Default isFalse, which treats unknown values as errors.

template_suffixOptional[str]

Treatname as a template table and provide a suffix. BigQuery will create the table based on the schema of the template table. Seehttps://cloud.google.com/bigquery/streaming-data-into-bigquery#template-tables

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
TypeErrorifjson_rows is not aSequence.
Returns
TypeDescription
Sequence[Mappings]One mapping per row with insert errors: the "index" key identifies the row, and the "errors" key contains a list of the mappings describing one or more problems with the row.

job_from_resource

job_from_resource(resource:dict,)->typing.Union[google.cloud.bigquery.job.copy_.CopyJob,google.cloud.bigquery.job.extract.ExtractJob,google.cloud.bigquery.job.load.LoadJob,google.cloud.bigquery.job.query.QueryJob,google.cloud.bigquery.job.base.UnknownJob,]

Detect correct job type from resource and instantiate.

Parameter
NameDescription
resourceDict

one job resource from API response

Returns
TypeDescription
Union[job.CopyJob, job.ExtractJob, job.LoadJob, job.QueryJob, job.UnknownJob]The job instance, constructed via the resource.

list_datasets

list_datasets(project:typing.Optional[str]=None,include_all:bool=False,filter:typing.Optional[str]=None,max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,page_size:typing.Optional[int]=None,)->google.api_core.page_iterator.Iterator

List datasets for the project associated with this client.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/datasets/list

Parameters
NameDescription
projectOptional[str]

Project ID to use for retreiving datasets. Defaults to the client's project.

include_allOptional[bool]

True if results include hidden datasets. Defaults to False.

filterOptional[str]

An expression for filtering the results by label. For syntax, seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/datasets/list#body.QUERY_PARAMETERS.filter

max_resultsOptional[int]

Maximum number of datasets to return.

page_tokenOptional[str]

Token representing a cursor into the datasets. If not passed, the API will return the first page of datasets. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of thegoogle.api_core.page_iterator.HTTPIterator.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

page_sizeOptional[int]

Maximum number of datasets to return per page.

Returns
TypeDescription
google.api_core.page_iterator.IteratorIterator ofDatasetListItem. associated with the project.

list_jobs

list_jobs(project:typing.Optional[str]=None,parent_job:typing.Optional[typing.Union[google.cloud.bigquery.job.query.QueryJob,str]]=None,max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,all_users:typing.Optional[bool]=None,state_filter:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,min_creation_time:typing.Optional[datetime.datetime]=None,max_creation_time:typing.Optional[datetime.datetime]=None,page_size:typing.Optional[int]=None,)->google.api_core.page_iterator.Iterator

List jobs for the project associated with this client.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/list

Parameters
NameDescription
projectOptional[str]

Project ID to use for retreiving datasets. Defaults to the client's project.

parent_jobOptional[Union[google.cloud.bigquery.job._AsyncJob, str, ]]

If set, retrieve only child jobs of the specified parent.

max_resultsOptional[int]

Maximum number of jobs to return.

page_tokenOptional[str]

Opaque marker for the next "page" of jobs. If not passed, the API will return the first page of jobs. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token ofgoogle.api_core.page_iterator.HTTPIterator.

all_usersOptional[bool]

If true, include jobs owned by all users in the project. Defaults to :data:False.

state_filterOptional[str]

If set, include only jobs matching the given state. One of: *"done" *"pending" *"running"

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

min_creation_timeOptional[datetime.datetime]

Min value for job creation time. If set, only jobs created after or at this timestamp are returned. If the datetime has no time zone assumes UTC time.

max_creation_timeOptional[datetime.datetime]

Max value for job creation time. If set, only jobs created before or at this timestamp are returned. If the datetime has no time zone assumes UTC time.

page_sizeOptional[int]

Maximum number of jobs to return per page.

Returns
TypeDescription
google.api_core.page_iterator.IteratorIterable of job instances.

list_models

list_models(dataset:typing.Union[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem,str,],max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,page_size:typing.Optional[int]=None,)->google.api_core.page_iterator.Iterator
Parameters
NameDescription
datasetUnion[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem, str, ]

A reference to the dataset whose models to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string usingfrom_string.

max_resultsOptional[int]

Maximum number of models to return. Defaults to a value set by the API.

page_tokenOptional[str]

Token representing a cursor into the models. If not passed, the API will return the first page of models. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of thegoogle.api_core.page_iterator.HTTPIterator.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

page_sizeOptional[int] Returns: google.api_core.page_iterator.Iterator: Iterator ofModel contained within the requested dataset.

Maximum number of models to return per page. Defaults to a value set by the API.

list_partitions

list_partitions(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->typing.Sequence[str]

List the partitions in a table.

Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

The table or reference from which to get partition info

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry. If multiple requests are made under the hood,timeout applies to each individual request.

Returns
TypeDescription
List[str]A list of the partition ids present in the partitioned table

list_projects

list_projects(max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,page_size:typing.Optional[int]=None,)->google.api_core.page_iterator.Iterator

List projects for the project associated with this client.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/projects/list

Parameters
NameDescription
max_resultsOptional[int]

Maximum number of projects to return. Defaults to a value set by the API.

page_tokenOptional[str]

Token representing a cursor into the projects. If not passed, the API will return the first page of projects. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of thegoogle.api_core.page_iterator.HTTPIterator.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

page_sizeOptional[int]

Maximum number of projects to return in each page. Defaults to a value set by the API.

Returns
TypeDescription
google.api_core.page_iterator.IteratorIterator ofProject accessible to the current client.

list_routines

list_routines(dataset:typing.Union[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem,str,],max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,page_size:typing.Optional[int]=None,)->google.api_core.page_iterator.Iterator
Parameters
NameDescription
datasetUnion[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem, str, ]

A reference to the dataset whose routines to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string usingfrom_string.

max_resultsOptional[int]

Maximum number of routines to return. Defaults to a value set by the API.

page_tokenOptional[str]

Token representing a cursor into the routines. If not passed, the API will return the first page of routines. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of thegoogle.api_core.page_iterator.HTTPIterator.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

page_sizeOptional[int] Returns: google.api_core.page_iterator.Iterator: Iterator of allRoutines contained within the requested dataset, limited bymax_results.

Maximum number of routines to return per page. Defaults to a value set by the API.

list_rows

list_rows(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableListItem,google.cloud.bigquery.table.TableReference,str,],selected_fields:typing.Optional[typing.Sequence[google.cloud.bigquery.schema.SchemaField]]=None,max_results:typing.Optional[int]=None,page_token:typing.Optional[str]=None,start_index:typing.Optional[int]=None,page_size:typing.Optional[int]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.table.RowIterator
Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableListItem,google.cloud.bigquery.table.TableReference, str, ]

The table to list, or a reference to it. When the table object does not contain a schema andselected_fields is not supplied, this method callsget_table to fetch the table schema.

selected_fieldsSequence[google.cloud.bigquery.schema.SchemaField]

The fields to return. If not supplied, data for all columns are downloaded.

max_resultsOptional[int]

Maximum number of rows to return.

page_tokenOptional[str]

Token representing a cursor into the table's rows. If not passed, the API will return the first page of the rows. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of theRowIterator.

start_indexOptional[int]

The zero-based index of the starting row to read.

page_sizeOptional[int]

The maximum number of rows in each page of results from this request. Non-positive values are ignored. Defaults to a sensible value set by the API.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry. If multiple requests are made under the hood,timeout applies to each individual request.

Returns
TypeDescription
google.cloud.bigquery.table.RowIteratorIterator of row dataRow-s. During each page, the iterator will have thetotal_rows attribute set, which counts the total number of rows **in the table** (this is distinct from the total number of rows in the current page:iterator.page.num_items).

list_tables

Parameters
NameDescription
datasetUnion[google.cloud.bigquery.dataset.Dataset,google.cloud.bigquery.dataset.DatasetReference,google.cloud.bigquery.dataset.DatasetListItem, str, ]

A reference to the dataset whose tables to list from the BigQuery API. If a string is passed in, this method attempts to create a dataset reference from a string usingfrom_string.

max_resultsOptional[int]

Maximum number of tables to return. Defaults to a value set by the API.

page_tokenOptional[str]

Token representing a cursor into the tables. If not passed, the API will return the first page of tables. The token marks the beginning of the iterator to be returned and the value of thepage_token can be accessed atnext_page_token of thegoogle.api_core.page_iterator.HTTPIterator.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

page_sizeOptional[int]

Maximum number of tables to return per page. Defaults to a value set by the API.

Returns
TypeDescription
google.api_core.page_iterator.IteratorIterator ofTableListItem contained within the requested dataset.

load_table_from_dataframe

load_table_from_dataframe(dataframe:pandas.DataFrame,destination:Union[Table,TableReference,str],num_retries:int=6,job_id:Optional[str]=None,job_id_prefix:Optional[str]=None,location:Optional[str]=None,project:Optional[str]=None,job_config:Optional[LoadJobConfig]=None,parquet_compression:str="snappy",timeout:ResumableTimeoutType=None,)->job.LoadJob
Parameters
NameDescription
dataframepandas.Dataframe

Apandas.DataFrame containing the data to load.

destinationUnion[ Table, TableReference, str ]

The destination table to use for loading the data. If it is an existing table, the schema of thepandas.DataFrame must match the schema of the destination table. If the table does not yet exist, the schema is inferred from thepandas.DataFrame. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

num_retriesOptional[int]

Number of upload retries. Defaults to 6.

job_idOptional[str]

Name of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[LoadJobConfig]

Extra configuration options for the job. To override the default pandas data type conversions, supply a value forschema with column names matching those of the dataframe. The BigQuery schema is used to determine the correct data type conversion. Indexes are not loaded. By default, this method uses the parquet source format. To override this, supply a value forsource_format with the format name. Currently onlyCSV andPARQUET are supported.

parquet_compressionOptional[str]

[Beta] The compression method to use if intermittently serializingdataframe to a parquet file. Defaults to "snappy". The argument is directly passed as thecompression argument to the underlyingpyarrow.parquet.write_table() method (the default value "snappy" gets converted to uppercase).https://arrow.apache.org/docs/python/generated/pyarrow.parquet.write_table.html#pyarrow-parquet-write-table If the job config schema is missing, the argument is directly passed as thecompression argument to the underlyingDataFrame.to_parquet() method.https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.to_parquet.html#pandas.DataFrame.to_parquet

timeoutOptional[flaot]

The number of seconds to wait for the underlying HTTP transport before usingretry. Depending on the retry strategy, a request may be repeated several times using the same timeout each time. Defaults to None. Can also be passed as a tuple (connect_timeout, read_timeout). Seerequests.Session.request documentation for details.

Exceptions
TypeDescription
ValueErrorIf a usable parquet engine cannot be found. This method requirespyarrow to be installed.
TypeErrorIfjob_config is not an instance ofLoadJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.job.LoadJobA new load job.

load_table_from_file

load_table_from_file(file_obj:typing.IO[bytes],destination:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],rewind:bool=False,size:typing.Optional[int]=None,num_retries:int=6,job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,job_config:typing.Optional[google.cloud.bigquery.job.load.LoadJobConfig]=None,timeout:typing.Union[None,float,typing.Tuple[float,float]]=None,)->google.cloud.bigquery.job.load.LoadJob

Upload the contents of this table from a file-like object.

Similar toload_table_from_uri, this method creates, starts andreturns a xref_LoadJob.

Parameters
NameDescription
file_objIO[bytes]

A file handle opened in binary mode for reading.

destinationUnion[Table, TableReference, TableListItem, str ]

Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

rewindOptional[bool]

If True, seek to the beginning of the file handle before reading the file. Defaults to False.

sizeOptional[int]

The number of bytes to read from the file handle. If size isNone or large, resumable upload will be used. Otherwise, multipart upload will be used.

num_retriesOptional[int]

Number of upload retries. Defaults to 6.

job_idOptional[str]

Name of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[LoadJobConfig]

Extra configuration options for the job.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry. Depending on the retry strategy, a request may be repeated several times using the same timeout each time. Defaults to None. Can also be passed as a tuple (connect_timeout, read_timeout). Seerequests.Session.request documentation for details.

Exceptions
TypeDescription
ValueErrorIfsize is not passed in and can not be determined, or if thefile_obj can be detected to be a file opened in text mode.
TypeErrorIfjob_config is not an instance ofLoadJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.job.LoadJobA new load job.

load_table_from_json

load_table_from_json(json_rows:typing.Iterable[typing.Dict[str,typing.Any]],destination:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],num_retries:int=6,job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,job_config:typing.Optional[google.cloud.bigquery.job.load.LoadJobConfig]=None,timeout:typing.Union[None,float,typing.Tuple[float,float]]=None,)->google.cloud.bigquery.job.load.LoadJob

Upload the contents of a table from a JSON string or dict.

Parameters
NameDescription
json_rowsIterable[Dict[str, Any]]

Row data to be inserted. Keys must match the table schema fields and values must be JSON-compatible representations. .. note:: If your data is already a newline-delimited JSON string, it is best to wrap it into a file-like object and pass it toload_table_from_file:: import io from google.cloud import bigquery data = u'{"foo": "bar"}' data_as_file = io.StringIO(data) client = bigquery.Client() client.load_table_from_file(data_as_file, ...)

destinationUnion[ Table, TableReference, TableListItem, str ]

Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

num_retriesOptional[int]

Number of upload retries. Defaults to 6.

job_idOptional[str]

Name of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[LoadJobConfig]

Extra configuration options for the job. Thesource_format setting is always set toNEWLINE_DELIMITED_JSON.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry. Depending on the retry strategy, a request may be repeated several times using the same timeout each time. Defaults to None. Can also be passed as a tuple (connect_timeout, read_timeout). Seerequests.Session.request documentation for details.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofLoadJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.job.LoadJobA new load job.

load_table_from_uri

load_table_from_uri(source_uris:typing.Union[str,typing.Sequence[str]],destination:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,job_config:typing.Optional[google.cloud.bigquery.job.load.LoadJobConfig]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.job.load.LoadJob

Starts a job for loading data into a table from Cloud Storage.

Seehttps://cloud.google.com/bigquery/docs/reference/rest/v2/Job#jobconfigurationload

Parameters
NameDescription
source_urisUnion[str, Sequence[str]]

URIs of data files to be loaded; in formatgs://<bucket_name>/<object_name_or_glob>.

destinationUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

Table into which data is to be loaded. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

job_idOptional[str]

Name of the job.

job_id_prefixOptional[str]

The user-provided prefix for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

job_configOptional[google.cloud.bigquery.job.LoadJobConfig]

Extra configuration options for the job.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofLoadJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.job.LoadJobA new load job.

query

query(query:str,job_config:typing.Optional[google.cloud.bigquery.job.query.QueryJobConfig]=None,job_id:typing.Optional[str]=None,job_id_prefix:typing.Optional[str]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,job_retry:typing.Optional[google.api_core.retry.retry_unary.Retry]=google.api_core.retry.retry_unary.Retry,api_method:typing.Union[str,google.cloud.bigquery.enums.QueryApiMethod]=QueryApiMethod.INSERT,)->google.cloud.bigquery.job.query.QueryJob
Parameters
NameDescription
querystr

SQL query to be executed. Defaults to the standard SQL dialect. Use thejob_config parameter to change dialects.

job_configOptional[google.cloud.bigquery.job.QueryJobConfig]

Extra configuration options for the job. To override any options that were previously set in thedefault_query_job_config given to theClient constructor, manually set those options toNone, or whatever value is preferred.

job_idOptional[str]

ID to use for the query job.

job_id_prefixOptional[str]

The prefix to use for a randomly generated job ID. This parameter will be ignored if ajob_id is also given.

locationOptional[str]

Location where to run the job. Must match the location of the table used in the query as well as the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC. This only applies to making RPC calls. It isn't used to retry failed jobs. This has a reasonable default that should only be overridden with care.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

job_retryOptional[google.api_core.retry.Retry]

How to retry failed jobs. The default retries rate-limit-exceeded errors. PassingNone disables job retry. Not all jobs can be retried. Ifjob_id is provided, then the job returned by the query will not be retryable, and an exception will be raised if a non-None (and non-default) value forjob_retry is also provided. Note that errors aren't detected untilresult() is called on the job returned. Thejob_retry specified here becomes the defaultjob_retry forresult(), where it can also be specified.

api_methodUnion[str, enums.QueryApiMethod]

Method with which to start the query job. By default, the jobs.insert API is used for starting a query. SeeQueryApiMethod for details on the difference between the query start methods.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofQueryJobConfig class, or if bothjob_id and non-None non-defaultjob_retry are provided.
Returns
TypeDescription
google.cloud.bigquery.job.QueryJobA new query job instance.

query_and_wait

query_and_wait(query,*,job_config:typing.Optional[google.cloud.bigquery.job.query.QueryJobConfig]=None,location:typing.Optional[str]=None,project:typing.Optional[str]=None,api_timeout:typing.Optional[float]=None,wait_timeout:typing.Union[float,None,object]=object,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,job_retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,page_size:typing.Optional[int]=None,max_results:typing.Optional[int]=None)->google.cloud.bigquery.table.RowIterator

Run the query, wait for it to finish, and return the results.

Parameters
NameDescription
querystr

SQL query to be executed. Defaults to the standard SQL dialect. Use thejob_config parameter to change dialects.

job_configOptional[google.cloud.bigquery.job.QueryJobConfig]

Extra configuration options for the job. To override any options that were previously set in thedefault_query_job_config given to theClient constructor, manually set those options toNone, or whatever value is preferred.

locationOptional[str]

Location where to run the job. Must match the location of the table used in the query as well as the destination table.

projectOptional[str]

Project ID of the project of where to run the job. Defaults to the client's project.

api_timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

wait_timeoutOptional[Union[float, object]]

The number of seconds to wait for the query to finish. If the query doesn't finish before this timeout, the client attempts to cancel the query. If unset, the underlying REST API calls have timeouts, but we still wait indefinitely for the job to finish.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC. This only applies to making RPC calls. It isn't used to retry failed jobs. This has a reasonable default that should only be overridden with care.

job_retryOptional[google.api_core.retry.Retry]

How to retry failed jobs. The default retries rate-limit-exceeded errors. PassingNone disables job retry. Not all jobs can be retried.

page_sizeOptional[int]

The maximum number of rows in each page of results from the initial jobs.query request. Non-positive values are ignored.

max_resultsOptional[int]

The maximum total number of rows from this request.

Exceptions
TypeDescription
TypeErrorIfjob_config is not an instance ofQueryJobConfig class.
Returns
TypeDescription
google.cloud.bigquery.table.RowIteratorIterator of row dataRow-s. During each page, the iterator will have thetotal_rows attribute set, which counts the total number of rows **in the result set** (this is distinct from the total number of rows in the current page:iterator.page.num_items). If the query is a special query that produces no results, e.g. a DDL query, an_EmptyRowIterator instance is returned.

schema_from_json

schema_from_json(file_or_path:PathType)->List[SchemaField]

Takes a file object or file path that contains json that describesa table schema.

Returns
TypeDescription
List[SchemaField]List ofSchemaField objects.

schema_to_json

schema_to_json(schema_list:Sequence[SchemaField],destination:PathType)

Takes a list of schema field objects.

Serializes the list of schema field objects as json to a file.

Destination is a file path or a file object.

set_iam_policy

set_iam_policy(table:typing.Union[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem,str,],policy:google.api_core.iam.Policy,updateMask:typing.Optional[str]=None,retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,*,fields:typing.Sequence[str]=())->google.api_core.iam.Policy

Return the access control policy for a table resource.

Parameters
NameDescription
tableUnion[google.cloud.bigquery.table.Table,google.cloud.bigquery.table.TableReference,google.cloud.bigquery.table.TableListItem, str, ]

The table to get the access control policy for. If a string is passed in, this method attempts to create a table reference from a string usingfrom_string.

policygoogle.api_core.iam.Policy

The access control policy to set.

updateMaskOptional[str]

Mask as defined byhttps://cloud.google.com/bigquery/docs/reference/rest/v2/tables/setIamPolicy#body.request_body.FIELDS.update_mask Incompatible withfields.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

fieldsSequence[str]

Which properties to set on the policy. See:https://cloud.google.com/bigquery/docs/reference/rest/v2/tables/setIamPolicy#body.request_body.FIELDS.update_mask Incompatible withupdateMask.

Returns
TypeDescription
google.api_core.iam.PolicyThe updated access control policy.

update_dataset

update_dataset(dataset:google.cloud.bigquery.dataset.Dataset,fields:typing.Sequence[str],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,update_mode:typing.Optional[google.cloud.bigquery.enums.UpdateMode]=None,)->google.cloud.bigquery.dataset.Dataset

Change some fields of a dataset.

Usefields to specify which fields to update. At least one fieldmust be provided. If a field is listed infields and isNone indataset, it will be deleted.

For example, to update the default expiration times, specifyboth properties in thefields argument:

bigquery_client.update_dataset(    dataset,    [        "default_partition_expiration_ms",        "default_table_expiration_ms",    ])

Ifdataset.etag is notNone, the update will onlysucceed if the dataset on the server has the same ETag. Thusreading a dataset withget_dataset, changing its fields,and then passing it toupdate_dataset will ensure that the changeswill only be saved if no modifications to the dataset occurredsince the read.

Parameters
NameDescription
datasetgoogle.cloud.bigquery.dataset.Dataset

The dataset to update.

fieldsSequence[str]

The properties ofdataset to change. These are strings corresponding to the properties ofDataset.

retryOptional[google.api_core.retry.Retry]

How to retry the RPC.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

update_modeOptional[google.cloud.bigquery.enums.UpdateMode]

Specifies the kind of information to update in a dataset. By default, dataset metadata (e.g. friendlyName, description, labels, etc) and ACL information are updated. This argument can take on the following possible enum values. *UPDATE_MODE_UNSPECIFIED: The default value. Behavior defaults to UPDATE_FULL. *UPDATE_METADATA: Includes metadata information for the dataset, such as friendlyName, description, labels, etc. *UPDATE_ACL: Includes ACL information for the dataset, which defines dataset access for one or more entities. *UPDATE_FULL: Includes both dataset metadata and ACL information.

Returns
TypeDescription
google.cloud.bigquery.dataset.DatasetThe modifiedDataset instance.

update_model

update_model(model:google.cloud.bigquery.model.Model,fields:typing.Sequence[str],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.model.Model

[Beta] Change some fields of a model.

Usefields to specify which fields to update. At least one fieldmust be provided. If a field is listed infields and isNoneinmodel, the field value will be deleted.

For example, to update the descriptive properties of the model,specify them in thefields argument:

bigquery_client.update_model(    model, ["description", "friendly_name"])

Ifmodel.etag is notNone, the update will only succeed ifthe model on the server has the same ETag. Thus reading a model withget_model, changing its fields, and then passing it toupdate_model will ensure that the changes will only be saved ifno modifications to the model occurred since the read.

Parameters
NameDescription
modelgoogle.cloud.bigquery.model.Model

The model to update.

fieldsSequence[str]

The properties ofmodel to change. These are strings corresponding to the properties ofModel.

retryOptional[google.api_core.retry.Retry]

A description of how to retry the API call.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.model.ModelThe model resource returned from the API call.

update_routine

update_routine(routine:google.cloud.bigquery.routine.routine.Routine,fields:typing.Sequence[str],retry:google.api_core.retry.retry_unary.Retry=google.api_core.retry.retry_unary.Retry,timeout:typing.Optional[float]=None,)->google.cloud.bigquery.routine.routine.Routine

[Beta] Change some fields of a routine.

Usefields to specify which fields to update. At least one fieldmust be provided. If a field is listed infields and isNoneinroutine, the field value will be deleted.

For example, to update the description property of the routine,specify it in thefields argument:

bigquery_client.update_routine(    routine, ["description"])

Warning:During beta, partial updates are not supported. You must provideall fields in the resource.If xref_etag is notNone, the update will only succeed if the resource on the serverhas the same ETag. Thus reading a routine withxref_get_routine, changingits fields, and then passing it to this method will ensure that thechanges will only be saved if no modifications to the resourceoccurred since the read.

Parameters
NameDescription
routinegoogle.cloud.bigquery.routine.Routine

The routine to update.

fieldsSequence[str]

The fields ofroutine to change, spelled as theRoutine properties.

retryOptional[google.api_core.retry.Retry]

A description of how to retry the API call.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.routine.RoutineThe routine resource returned from the API call.

update_table

Parameters
NameDescription
tablegoogle.cloud.bigquery.table.Table

The table to update.

fieldsSequence[str]

The fields oftable to change, spelled as theTable properties.

autodetect_schemabool

Specifies if the schema of the table should be autodetected when updating the table from the underlying source. Only applicable for external tables.

retryOptional[google.api_core.retry.Retry]

A description of how to retry the API call.

timeoutOptional[float]

The number of seconds to wait for the underlying HTTP transport before usingretry.

Returns
TypeDescription
google.cloud.bigquery.table.TableThe table resource returned from the API call.

Except as otherwise noted, the content of this page is licensed under theCreative Commons Attribution 4.0 License, and code samples are licensed under theApache 2.0 License. For details, see theGoogle Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025-12-16 UTC.