With MadKudu + BigQuery, send your engagement, intent or CRM data to MadKudu.
Note: this integration is in Beta, tell us what you think at product@madkudu.com
MadKudu Tip
If you'd like to send MadKudu data to your BigQuery (Destination), you can do so through the Amazon S3 Destination.
What does this integration do?
Turning on the BigQuery integration allows you to send BigQuery data to MadKudu: product usage, web visits, contact properties, account properties... etc to surface to your Sales reps in the MadKudu iFrame in Salesforce or leverage in a scoring model.
Important notes
Data in BigQuery typically resides in tables. When sharing data with MadKudu, the best practice is to create separate tables for Event data, Contact data. You don't need to send all objects: if you have your CRM already connected to MadKudu you can send only event data with the contact email for example.
In each table shared with MadKudu, there must be a column corresponding to a
created_at
andupdated_at
: MadKudu pulls data incrementallynew rows based on
created_at
and updated rows based on
updated_at
instead of your whole table at every pull to go faster
The integration uses the following BigQuery scope: https://www.googleapis.com/auth/bigquery which corresponds to View and manage your data in Google BigQuery and see the email address for your Google Account. If you'd like to restrict access to MadKudu to only certain dataset and tables, please configure these permissions in BigQuery to the user you'll connect with MadKudu.
What Objects can you send to MadKudu?
MadKudu works with the following entities and required properties (if the naming doesn't match exactly these column you'll be able to map them in MadKudu but they need to be in there):
Contacts
Column | Type | example | Comment |
id | VARCHAR(256) | 93d8AB4f69fa80 | Required |
VARCHAR(256) | john@slack.com | Required | |
created_at | TIMESTAMP | 2023-12-02 20:23:12.010000 | Required. UTC |
updated_at | TIMESTAMP | 2023-12-02 20:23:12.010000 | Required. UTC |
contact_{property} | depends on the property can be numeric or varchar or boolean |
| Optional. |
Events
Column | Type | example | Comment |
id | VARCHAR(256) | 03492f12-061b-4f54-b866-764b4f69fa80 | Required |
event_name | VARCHAR(256) | logged_in | Required |
event_timestamp | TIMESTAMP | 2023-12-02 20:22:51.780000 | Required |
contact_id | VARCHAR(256) | john@slack.com | Required (or send contact_id and the Contacts table) |
created_at | TIMESTAMP | 2023-12-02 20:23:12.010000 | Required |
updated_at | TIMESTAMP | 2023-12-02 20:23:12.010000 | Required |
event_{property} | depends on the property can be numeric or varchar or boolean |
| Optional |
How to send Bigquery data to MadKudu?
Pre-requisite
You have access to your BigQuery account and to MadKudu app
For MadKudu to start pulling data from your BigQuery account, please follow the following steps
Log in to MadKudu app (app.madkudu.com)
Click on Integrations > BigQuery
Click on "Grant access to Google BigQuery" and follow the authentication flow
Once connected you'll see a form asking for
Project ID: located in your project list in your Google Console
Dataset: the name of your dataset
Location: the location of your dataset
Enter the table names corresponding to Events, Contacts or Accounts (coming soon) (you don't need to send MadKudu all 3, only the ones needed for your use case)
For each object, map your BigQuery table column names corresponding to the required properties. In the example below, the column in BigQuery corresponding to the event time is named "timestamp" and needs to be mapped into the event_timestamp (MadKudu naming)
Click Save
Click on Test connection to check MadKudu can connect to your dataset
MadKudu will start pulling data moving forward! For MadKudu to pull historical data on these tables, please submit a support ticket - it's a quick configuration to turn on on our end.
Example of request:
"Hi, we'd like to pull all data from our BigQuery instance connected to MadKudu
Table Events: from YYYY-MM-DD (we recommend 9 months of data)
Table Contacts: from YYYY-MM-DD
Table Account: from YYYY-MM-DD (coming soon)
Thanks"
At any time, you can see the status of the integration in Settings > Processes, under Connectors - Pull you should see the BigQuery integration and any error.
After a few hours, you can see the volume of Data already pulled by going to Data in the main left navigation bar and BigQuery
FAQ
Which BigQuery scope does this integration use when authenticating?
The integration uses the following BigQuery scope: https://www.googleapis.com/auth/bigquery which corresponds to View and manage your data in Google BigQuery and see the email address for your Google Account.