Source code for bigframes.bigquery._operations.io

# Copyright 2026 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from __future__ import annotations

from typing import Mapping, Optional, Union

import pandas as pd

from bigframes.bigquery._operations.table import _get_table_metadata
import bigframes.core.logging.log_adapter as log_adapter
import bigframes.core.sql.io
import bigframes.session


[docs] @log_adapter.method_logger(custom_base_name="bigquery_io") def load_data( table_name: str, *, write_disposition: str = "INTO", columns: Optional[Mapping[str, str]] = None, partition_by: Optional[list[str]] = None, cluster_by: Optional[list[str]] = None, table_options: Optional[Mapping[str, Union[str, int, float, bool, list]]] = None, from_files_options: Mapping[str, Union[str, int, float, bool, list]], with_partition_columns: Optional[Mapping[str, str]] = None, connection_name: Optional[str] = None, session: Optional[bigframes.session.Session] = None, ) -> pd.Series: """ Loads data into a BigQuery table. See the `BigQuery LOAD DATA DDL syntax <https://docs.cloud.google.com/bigquery/docs/reference/standard-sql/load-statements>`_ for additional reference. Args: table_name (str): The name of the table in BigQuery. write_disposition (str, default "INTO"): Whether to replace the table if it already exists ("OVERWRITE") or append to it ("INTO"). columns (Mapping[str, str], optional): The table's schema. partition_by (list[str], optional): A list of partition expressions to partition the table by. See https://docs.cloud.google.com/bigquery/docs/reference/standard-sql/load-statements#partition_expression. cluster_by (list[str], optional): A list of columns to cluster the table by. table_options (Mapping[str, Union[str, int, float, bool, list]], optional): The table options. from_files_options (Mapping[str, Union[str, int, float, bool, list]]): The options for loading data from files. with_partition_columns (Mapping[str, str], optional): The table's partition columns. connection_name (str, optional): The connection to use for the table. session (bigframes.session.Session, optional): The session to use. If not provided, the default session is used. Returns: pandas.Series: A Series with object dtype containing the table metadata. Reference the `BigQuery Table REST API reference <https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#Table>`_ for available fields. """ import bigframes.pandas as bpd sql = bigframes.core.sql.io.load_data_ddl( table_name=table_name, write_disposition=write_disposition, columns=columns, partition_by=partition_by, cluster_by=cluster_by, table_options=table_options, from_files_options=from_files_options, with_partition_columns=with_partition_columns, connection_name=connection_name, ) if session is None: bpd.read_gbq_query(sql) session = bpd.get_global_session() else: session.read_gbq_query(sql) return _get_table_metadata(bqclient=session.bqclient, table_name=table_name)