This post helps you with loading your data from Search Console to Snowflake. If you are looking to get analytics-ready data without the manual hassle you can integrate Search Console to Snowflake with Blendo, so you can focus on what matters, getting value out of your Google Search related data.
ACCESS YOUR DATA ON GOOGLE SEARCH CONSOLE
The first step in loading your search Console data to any kind of data warehouse solution, is to access your data and start extracting it.
You access your data for the Google Search Console through the Search Console APIs. There are two APIs available there,
- Search Console API
- URL Testing Tools API
From the two, we are interested in the first API which allows us to access any data we are interested for.
As every other Google product, you need to authorize yourself to get access to the API through an implementation of the OAuth 2.0 protocol. The API is web based following a REST like architecture but Google also offers some SDKs that you can use for some popular languages like Java and Python.
The things that you have to keep in mind when dealing with any API like the one the Google Search Console has, are:
- Rate limits. Every API, has some rate limits that you have to respect.
- Authentication. You authenticate on Google using an OAuth.
- Paging and dealing with big amount of data. Platforms like Google tend to generate a lot of data. Pulling big volumes of data out of an API might be difficult, especially when you consider and respect any rate limits that the API has.
ABOUT GOOGLE SEARCH CONSOLE
Google Search Console is a product offered by Google to web administrators. It allows you to submit sitemaps to Google, trigger the indexing of your website and see statistics about what’s going on, like possible errors and speed related problems.
Most importantly, Google Search Console offers a wealth of statistics about the queries that users are performing in order to click on a link and get on one of your landing pages. This information can help tremendously in search engine optimization and when you are serious about content marketing.
You need to have in mind the following about Google Search Console.
- You see only sample data, and
- You can get up to 90 days of data
So, it’s important to start collecting and storing your Search Console data as soon as possible and make sure that you sync all the available data.
TRANSFORM AND PREPARE YOUR GOOGLE SEARCH CONSOLE DATA FOR SNOWFLAKE
After you have accessed your data on Google Search Console, you will have to transform it based on two main factors,
- The limitations of the database that is going to be used
- The type of analysis that you plan to perform
Each system has specific limitations on the data types and data structures that it supports. If for example you want to push data into Google BigQuery, then you can send nested data like JSON directly. But when you are dealing with tabular data stores, like PostgreSQL, this is not an option. Instead, you will have to flatten out your data before loading into the database.
Also, you have to choose the right data types. Again, depending on the system that you will send data to and the data types that the API exposes to you, you will have to make the right choices. These choices are important because they can limit the expressivity of your queries and limit your analysts on what they can do directly out of the database.
Google Search Console data is modeled around the concept of a report, just like Google Analytics but with a much more limited number of dimensions and metrics.
At the end you will need to map one report to a table on your database and make sure that all data is stored into it. Dimensions and metrics will become columns of the tables.
You need to take special care of the fact that the reports you will be getting from Google Search Console, do not have primary keys given by Google, in order to avoid duplicates.
For more information on how you can query your Search Analytics data, please see here.
Data in Snowflake is organized around tables with a well-defined set of columns with each one having a specific data type.
Snowflake supports a rich set of data types. It is worth mentioning that a number of semi-structured data types is also supported. With Snowflake, is possible to load directly data in JSON, Avro, ORC, Parquet, or XML format. Hierarchical data is treated as a first-class citizen, similar to what Google BigQuery offers.
There is also one notable common data type that is not supported by Snowflake. LOB or large object data type is not supported, instead, you should use a BINARY or VARCHAR type instead. But these types are not that useful for data warehouse use cases.
A typical strategy for loading data from Google Search Console to Snowflake is to create a schema where you will map each API endpoint to a table.
Each key inside the Google Search Console API endpoint response should be mapped to a column of that table and you should ensure the right conversion to a Snowflake data type.
Of course, you will need to ensure that as data types from the Google Search Console API might change, you will adapt your database tables accordingly, there’s no such thing as automatic data type casting.
After you have a complete and well-defined data model or schema for Snowflake, you can move forward and start loading your data into the database.
LOAD DATA FROM GOOGLE SEARCH CONSOLE TO SNOWFLAKE
Usually, data is loaded into Snowflake in a bulk way, using the COPY INTO command. Files containing data, usually in JSON format, are stored in a local file system or in Amazon S3 buckets. Then a COPY INTO command is invoked on the Snowflake instance and data is copied into a data warehouse.
The files can be pushed into Snowflake using the PUT command, into a staging environment before the COPY command is invoked.
Another alternative is to upload every data directly into a service like Amazon S3 from where Snowflake can access data directly.
UPDATING YOUR GOOGLE SEARCH CONSOLE DATA ON SNOWFLAKE
As you will be generating more data on Google Search Console, you will need to update your older data on Snowflake. This includes new records together with updates to older records that for any reason have been updated on Google Search Console.
You will need to periodically check Google Search Console for new data and repeat the process that has been described previously while updating your currently available data if needed. Updating an already existing row on a Snowflake table is achieved by creating UPDATE statements.
Another issue that you need to take care of is the identification and removal of any duplicate records on your database. Either because Google Search Console does not have a mechanism to identify new and updated records or because of errors on your data pipelines, duplicate records might be introduced to your database.
In general, ensuring the quality of data that is inserted into your database is a big and difficult issue.
The best way to load data from Google Search Console to Snowflake
So far we just scraped the surface of what you can do with Snowflake and how to load data into it. Things can get even more complicated if you want to integrate data coming from different sources.
Are you striving to achieve results right now?
Instead of writing, hosting and maintaining a flexible data infrastructure use Blendo that can handle everything automatically for you.
Blendo with one click integrates with sources or services, creates analytics-ready data and syncs your Search Console to Snowflake right away.