BigQuery BI Engine is a fast, in-memory analysis service. By using BI Engine you can analyze data stored in BigQuery with sub-second query response time and with high concurrency.
BI Engine integrates with familiar Google tools like Google Data Studio to accelerate data exploration and analysis. With BI Engine, you can build rich, interactive dashboards and reports in Data Studio without compromising performance, scale, security, or data freshness.
https://cloud.google.com/solutions/data-warehouse-modernization
Streamline your migration path and unlock intelligent insights with BigQuery
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-migration-overview
What and how to migrate: The migration framework
Transition state
Fully migrated
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-schema-and-data-transfer-overview
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-data-governance
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-data-pipelines
This document is part of a series that explores how to migrate your upstream data pipelines, which load data into your data warehouse. This document discusses data pipelines: what they are and what to think about when migrating them.
In the context of data warehousing, data pipelines are commonly used to read data from transactional systems, apply transformations, and then write data to the data warehouse. In the context of data pipelines, sources are usually transactional systems—for example, an RDBMS—and the sink connects to a data warehouse. This type of graph is referred to as a data flow DAG. You can also use DAGs to orchestrate data movement between data pipelines and other systems. This usage is referred to as an orchestration or control flow DAG.
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-reporting-and-analysis
https://cloud.google.com/solutions/migration/dw2bq/dw-bq-performance-optimization
Use the Power of Google Cloud and Informatica to Build a Modern Data Architecture (Cloud Next '18)
Exploring and Preparing your Data with BigQuery https://www.coursera.org/learn/gcp-exploring-preparing-data-bigquery
Creating New BigQuery Datasets and Visualizing Insights https://www.coursera.org/learn/gcp-creating-bigquery-datasets-visualizing-insights
Modernizing Data Lakes and Data Warehouses with GCP https://www.coursera.org/learn/data-lakes-data-warehouses-gcp/home/welcome
The two key components of any data pipeline are data lakes and warehouses. This course highlights use-cases for each type of storage and dives into the available data lake and warehouse solutions on Google Cloud Platform in technical detail. Also, this course describes the role of a data engineer, the benefits of a successful data pipeline to business operations, and examines why data engineering should be done in a cloud environment.
Sandbox https://cloud.google.com/bigquery/docs/sandbox
The BigQuery sandbox gives you free access to the power of BigQuery subject to the sandbox's limits. The sandbox allows you to use the web UI in the GCP Console without providing a credit card. You can use the sandbox without creating a billing account or enabling billing for your project.
15 Awesome things you probably didn’t know about Google BigQuery https://medium.com/google-cloud/15-awesome-things-you-probably-didnt-know-about-google-bigquery-6654841fa2dc
Access control examples https://cloud.google.com/bigquery/docs/access-control-examples
Creating authorized views https://cloud.google.com/bigquery/docs/authorized-views
Introduction to interacting with BigQuery https://cloud.google.com/bigquery/docs/interacting-with-bigquery
Provides an overview of ways to interact with BigQuery.
Introduction to partitioned tables Using GROUP BY to avoid self-joins
A partitioned table is a special table that is divided into segments, called partitions, that make it easier to manage and query your data. By dividing a large table into smaller partitions, you can improve query performance, and you can control costs by reducing the number of bytes read by a query.
Standard SQL Functions & Operators. https://cloud.google.com/bigquery/docs/reference/standard-sql/functions-and-operators
Let’s talk data. https://medium.com/@hoffa
Why Nesting Is So Cool https://looker.com/blog/why-nesting-is-so-cool
When you're setting up a data warehouse, one of the key questions is how to structure your data. Do you completely normalize the data into a snowflake schema? Completely denormalize it into a very wide table with lots of repeated values? Or do something in the middle like a star schema?
Exporting Table Data https://cloud.google.com/bigquery/docs/exporting-data
BigQuery can export up to 1 GB of data to a single file. If you are exporting more than 1 GB of data, you must export your data to multiple files.
You cannot export table data to a local file, to Google Sheets, or to Google Drive. The only supported export location is Cloud Storage.
Query plan and timeline https://cloud.google.com/bigquery/query-plan-explanation
Embedded within query jobs, BigQuery includes diagnostic query plan and timing information.
When BigQuery executes a query job, it converts the declarative SQL statement into a graph of execution, broken up into a series of query stages, which themselves are composed of more granular sets of execution steps. BigQuery leverages a heavily distributed parallel architecture to run these queries. Stages model the units of work that many potential workers may execute in parallel.
In-memory query execution in Google BigQuery https://cloud.google.com/blog/products/gcp/in-memory-query-execution-in-google-bigquery
How To Control Access To BigQuery At Row Level With Groups https://medium.com/google-cloud/how-to-control-access-to-bigquery-at-row-level-with-groups-1cbccb111d9e
https://cloud.google.com/blog/products/gcp/life-of-a-bigquery-streaming-insert
Google BigQuery, the fully managed cloud-native data warehouse for Google Cloud Platform (GCP) customers, supports several ways to ingest data into its managed storage, including explicit load jobs or via queries against external sources. These methods share a common theme, in that they're used to transfer and append this new storage as part of a (potentially large) single commit to a table. BigQuery also supports a method of ingestion known as streaming, which is intended to service the needs of users who need a more open-ended, continuous style of ingestion. In this post, you’ll learn how the streaming system works, which should help aid understanding of observed behaviors and make it easier to reason about building integrations with BigQuery.