Is there a way to access Google My Business Data in BigQuery? - etl

We are working on a project where we want to access all the information from our numerous Google My Business locations. We would like to access that GMB information with our Google Analytics data and have it automatically refresh for a dashboard. We have looked into using Super Metrics and loading the data in that way, but we want to avoid upgrading our Super Metrics account.
Has anyone encountered a similar issue, and if so, what solution did/would you implement?

I think you have two solutions here :
Pay supermetrics or any self managed data loader (stitch, fivetran etc)
Code your own ETL

For the moment paying supermetrics, but some alternatives we are looking at are Keboola, Dataddo or building into airbyte, may it will come soon.

Related

Using custom metrics in self-hosted sentry

I have started using sentry within my org and loving it so far.
I've been trying to use its performance monitoring tool with custom metrics added.
While I can add custom metrics to the transactions I'm generating in sentry_sdk (for Python), I can't get access to them on the dashboard of our self-hosted installation of sentry.
After a lot of digging, I came across this paragraph here which states that
This feature is only available to organization on our latest plans which include Dynamic Sampling. Customers on legacy plans must move to one of these plans in order to access custom metrics.
From what I gather, I believe their plans in general is to run sentry on their servers. Unless you opt-in to their self-hosted code that can be downloaded from github here.
This is absolutely a bummer because I know my org will not consider moving internal data to third-party servers.
Wondering if someone knows of a solution to this problem. If sentry folks know of (paid) options that enables this feature on self-hosted version or if someone has hacked into their open source code?
I'd also love to hear any out-of-the-box suggestion you folks might have.

SalesForce Commerce Cloud. Intercept data of users and orders

I'm not so familiar with Commerce Cloud product but I need to know one point and I hope community can help me.
I need to implement a feature for customer who use SF Commerce Cloud and I would like to know it is possible or not. Customer wants to send some of data such as orders and users to an additional storage. This is requirement of local law and they have to implement it to do business.
Is it possible to intercept some actions like order placing, modifying, deleting and e-store customers creation, modifying, deleting? It would be great if you help me with direction where I can find additional information because after several attempts I can't get access to trial version of Commerce Cloud.
Thank you!
Yes, it is possible to do this in various ways. One way might be to implement a Javascript tracking integration that runs in the customer's browser and is referenced by the Storefront application that is running on SFCC. Another way would be to implement what is known as an Integration Cartridge which would implement several export jobs and/or service connections to your third party storage solution.
There is no trial version of the platform. In order to access an instance for development purposes, you will need to work through your customer's sandbox instances or become a Salesforce Partner.
Please review the Getting Started documentation. See also: Demandware/SFCC prerequisites

Is there any way to show the Google Cloud Build Data on a separate Dashboard?

I am using Google Cloud build for auto-build process. I want to create a dashboard which shows the trigger details and logs on which I worked on.
I know about Stack Driver but need suggestions other than that.
You could use Google's APIs to list or describe the builds and also to get the information you want about the triggers. The Viewing build results documentation explains what kind of information you can retrieve, but I always suggest playing around with Google's API Explorer.

How to set local database for google safe browsing update API (v4)?

I am building a service for checking for phishing or malware URLs for one of my applications. This service will be running on google app engine. Now, I want to use google safe browsing's Update API (v4) to have local database of URL hashes. But I am having hard time to understand the setup process of the local database they have mentioned.
https://developers.google.com/safe-browsing/v4/local-databases
They do provide a Go source code to do something of this sort but its not descriptive enough to have my own implementation.
I want to setup the db on google cloud itself. Can anyone point me to any good documentation or some ways to do the same if you have tried this before.
you can use:
"https://safebrowsing.googleapis.com/v4/threatListUpdates:fetch?key=" + apiKey;
to download hash version database.
you may need send a post request to get the database according to https://developers.google.com/safe-browsing/v4/reference/rest/v4/threatListUpdates/fetch

Give collaborator access to analytics only

We would like to be able to provide a collaborator with access to our analytics only. Is this possible?
It's quite important for our marketing and business development colleagues to be able to access analytics, while at the same time not have access to the data browser and cloud code.
It is not possible currently on Parse.
You should raise this as a feature request on Parse developer forums https://groups.google.com/forum/#!forum/parse-developers

Resources