Front-end Google BigQuery with an ASP.net 4.5 application - Part 1
__________________
This three part article shows how to set up a Google BigQuery project, how to front-end that project with a sample ASP.net application available for download as a GitHub repository, and how the engineering behind that sample application works.
Part one here explains how to set up a Google Cloud Platform account, and how to build a basic BigQuery project.
Part two will first describe the sample application user experience, and will then focus on application configuration and the security set-up.
Part three will explain the engineering behind the application.
I want to say thank you to Eyal Peled, of Google, for taking time to answer some questions I had about BigQuery.
__________________
Google Cloud Platform offers a wide range of products that move computing from an ownership model to a rental model. In the Platform product line, Google BigQuery handles large data volumes and queries. Although BigQuery lacks some features of products like Oracle or SQL Server, it's perfect for many of the problems traditionally solved by these products, without the hassles. It avoids the hardware and dedicated DBA overhead, at prices competitive with more "established" RDBMS tools. Part one here explains how to set up and load data into a Google BigQuery account.
This three part article shows how to set up a Google BigQuery project, how to front-end that project with a sample ASP.net application available for download as a GitHub repository, and how the engineering behind that sample application works.
Part one here explains how to set up a Google Cloud Platform account, and how to build a basic BigQuery project.
Part two will first describe the sample application user experience, and will then focus on application configuration and the security set-up.
Part three will explain the engineering behind the application.
I want to say thank you to Eyal Peled, of Google, for taking time to answer some questions I had about BigQuery.
__________________
Google Cloud Platform offers a wide range of products that move computing from an ownership model to a rental model. In the Platform product line, Google BigQuery handles large data volumes and queries. Although BigQuery lacks some features of products like Oracle or SQL Server, it's perfect for many of the problems traditionally solved by these products, without the hassles. It avoids the hardware and dedicated DBA overhead, at prices competitive with more "established" RDBMS tools. Part one here explains how to set up and load data into a Google BigQuery account.
Relational database products traditionally group tables, stored procedures, etc. in "databases." In contrast, BigQuery groups tables in a user-defined "dataset" within a user-defined project. Note that a dataset has only tables, not the other machinery RDBMS products provide. The ASP.net application we'll explore in Part 2 uses this dataset for its queries. While BigQuery provides some sample datasets, I will show how to load a sample CSV file into the dataset. Note that for user-sourced data, BigQuery requires project billing activation.
I started at the Google BigQuery page and I signed in with a Google account. Just above the black up arrow towards the top of the page I clicked "Go To My Console".
1. Set up Google Cloud Platform - first page |
To set up the BigQuery project for this article, I clicked "CREATE PROJECT" in the Console
to open this window:
2. Create Project |
3. Set up the new BigQuery project |
4. Project Dashboard |
5. Add one or more Google APIs to the Google Cloud Platform project |
6. Set up project billing |
Next, I loaded a data file into
BigQueryDemoApp. BigQuery accepts CSV or JSON files at this step.
I used zip file "Complete ZIP Code Totals File" from the U.S.Census Bureau for the data. It unzips into CSV file "zbp11totals.txt". This 3 meg CSV file has
about 39 K rows - enough to show the basics of BigQuery without hitting the structural limits or price points. At the Project Dashboard, I clicked "BigQuery" to the left of the black arrow here:
I clicked the indicated blue down arrow here:
8. Pick the BigQueryDemoApp project |
9. Drill down to the BigQueryDemoAppDS dataset |
name:string, age:integer
and fortunately, the Census
Bureau provided the source ZIP Code Totals data file layout here.
This layout maps to a format compatible with BigQuery:
ZIP:string, NAME:string, EMPFLAG:string,
EMP_NF:string, EMP:integer, QP1_NF:string, QP1:integer, AP_NF:string,
AP:integer, EST:integer, CITY:string, STABBR:string, CTY_NAME:string
I proceeded through the Create and Import steps, and at Advanced
Options as shown here I set the "Header rows to skip" value to 1.
This completed the Google Cloud account and BigQuery project set-up. Part two will cover the sample application that front-ends that project.