이 콘텐츠는 선택한 언어로 제공되지 않습니다.

Chapter 114. Google BigQuery Component


Available as of Camel version 2.20

114.1. Component Description

The Google Bigquery component provides access to Cloud BigQuery Infrastructure via the Google Client Services API.

The current implementation does not use gRPC.

The current implementation does not support querying BigQuery i.e. is a producer only.

Maven users will need to add the following dependency to their pom.xml for this component:

<dependency>
    <groupId>org.apache.camel</groupId>
    <artifactId>camel-google-bigquery</artifactId>
    <version>x.x.x</version>
    <!-- use the same version as your Camel core version -->
</dependency>

114.2. Authentication Configuration

Google BigQuery component authentication is targeted for use with the GCP Service Accounts. For more information please refer to Google Cloud Platform Auth Guide

Google security credentials can be set explicitly via one of the two options:

  • Service Account Email and Service Account Key (PEM format)
  • GCP credentials file location

If both are set, the Service Account Email/Key will take precedence.

Or implicitly, where the connection factory falls back on Application Default Credentials.

OBS! The location of the default credentials file is configurable - via GOOGLE_APPLICATION_CREDENTIALS environment variable.

Service Account Email and Service Account Key can be found in the GCP JSON credentials file as client_email and private_key respectively.

114.3. URI Format

        google-bigquery://project-id:datasetId[:tableId]?[options]

114.4. Options

The Google BigQuery component supports 4 options which are listed below.

NameDescriptionDefaultType

projectId (producer)

Google Cloud Project Id

 

String

datasetId (producer)

BigQuery Dataset Id

 

String

connectionFactory (producer)

ConnectionFactory to obtain connection to Bigquery Service. If non provided the default one will be used

 

GoogleBigQuery ConnectionFactory

resolveProperty Placeholders (advanced)

Whether the component should resolve property placeholders on itself when starting. Only properties which are of String type can use property placeholders.

true

boolean

The Google BigQuery endpoint is configured using URI syntax:

google-bigquery:projectId:datasetId:tableName

with the following path and query parameters:

114.4.1. Path Parameters (3 parameters):

NameDescriptionDefaultType

projectId

Required Google Cloud Project Id

 

String

datasetId

Required BigQuery Dataset Id

 

String

tableId

BigQuery table id

 

String

114.4.2. Query Parameters (3 parameters):

NameDescriptionDefaultType

connectionFactory (producer)

ConnectionFactory to obtain connection to Bigquery Service. If non provided the default will be used.

 

GoogleBigQuery ConnectionFactory

useAsInsertId (producer)

Field name to use as insert id

 

String

synchronous (advanced)

Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported).

false

boolean

114.5. Message Headers

NameTypeDescription

CamelGoogleBigQuery.TableSuffix

String

Table suffix to use when inserting data

CamelGoogleBigQuery.InsertId

String

InsertId to use when inserting data

CamelGoogleBigQuery.PartitionDecorator

String

Partition decorator to indicate partition to use when inserting data

CamelGoogleBigQuery.TableId

String

Table id where data will be submitted. If specified will override endpoint configuration

114.6. Producer Endpoints

Producer endpoints can accept and deliver to BigQuery individual and grouped exchanges alike. Grouped exchanges have Exchange.GROUPED_EXCHANGE property set.

Goole BigQuery producer will send a grouped exchange in a single api call unless different table suffix or partition decorators are specified in which case it will break it down to ensure data is written with the correct suffix or partition decorator.

Google BigQuery endpoint expects the payload to be either a map or list of maps. A payload containing a map will insert a single row and a payload containing a list of map’s will insert a row for each entry in the list.

114.7. Template tables

Reference: https://cloud.google.com/bigquery/streaming-data-into-bigquery#template-tables

Templated tables can be specified using the GoogleBigQueryConstants.TABLE_SUFFIX header.

I.e. the following route will create tables and insert records sharded on a per day basis:

from("direct:start")
.header(GoogleBigQueryConstants.TABLE_SUFFIX, "_${date:now:yyyyMMdd}")
.to("google-bigquery:sampleDataset:sampleTable")

Note it is recommended to use partitioning for this use case.

114.8. Partitioning

Reference: https://cloud.google.com/bigquery/docs/creating-partitioned-tables

Partitioning is specified when creating a table and if set data will be automatically partitioned into separate tables. When inserting data a specific partition can be specified by setting the GoogleBigQueryConstants.PARTITION_DECORATOR header on the exchange.

114.9. Ensuring data consistency

Reference: https://cloud.google.com/bigquery/streaming-data-into-bigquery#dataconsistency

A insert id can be set on the exchange with the header GoogleBigQueryConstants.INSERT_ID or by specifying query parameter useAsInsertId. As an insert id need to be specified per row inserted the exchange header can’t be used when the payload is a list - if the payload is a list the GoogleBigQueryConstants.INSERT_ID will be ignored. In that case use the query parameter useAsInsertId.

Red Hat logoGithubRedditYoutubeTwitter

자세한 정보

평가판, 구매 및 판매

커뮤니티

Red Hat 문서 정보

Red Hat을 사용하는 고객은 신뢰할 수 있는 콘텐츠가 포함된 제품과 서비스를 통해 혁신하고 목표를 달성할 수 있습니다.

보다 포괄적 수용을 위한 오픈 소스 용어 교체

Red Hat은 코드, 문서, 웹 속성에서 문제가 있는 언어를 교체하기 위해 최선을 다하고 있습니다. 자세한 내용은 다음을 참조하세요.Red Hat 블로그.

Red Hat 소개

Red Hat은 기업이 핵심 데이터 센터에서 네트워크 에지에 이르기까지 플랫폼과 환경 전반에서 더 쉽게 작업할 수 있도록 강화된 솔루션을 제공합니다.

© 2024 Red Hat, Inc.