On May 15th, 2017, three metrics will be removed from the Reporting API of the Google Apps Admin SDK:
num_docs_internally_visible
num_docs_externally_visible
num_docs_shared_outside_domain
I use all of these metrics in a scripts that performs some audits of our G Suite domain.
The migration docs say to use num_owned_items_with_visibility_shared_externally_delta instead of num_docs_shared_outside_domain for instance, but I don't understand how a delta metric can be used as a replacement unless you keep track of the actual number from day zero on.
How do I get the number of externally shared documents as a total, not a delta value?
Based from the documentation given, num_owned_items_with_visibility_shared_externally_delta is the number of items within the user's domain account that are not public or visible to anyone with link, but shared explicitly either with users or groups outside the domain up to the date of the report. So I think you can use it to get the number of externally shared documents. Also it is stated in the notice that differences in metrics calculations exist in these changes, so review the notes as well as Key issues for relevant details.
Related
I'm curious if there is a way to see how many requests are made per API key in my Google Cloud Platform account. First, there was a number specified for the last 30 days at every key.
How to see which keys aren't used anymore, or which make countless requests?
Thanks in advance!
In order for you to view request count per API key, you may use Cloud Monitoring.
First, in your google cloud console's navigation menu, go to Monitoring > Metrics Explorer.
In the Resource Type, select "Consumed API" in order to view the APIs used by the account.
In the Metric, select "Request Count".
In the Group by, select "credential_id" to group the results by API Key.
In the Minimum alignment period, select "Custom".
In the Period Value, input "30".
In the Period Unit, select "days".
IMPORTANT NOTES:
You may change the Period Unit based on your desired period unit but please note that the Period Value must follow the change on Period Unit.
You may change the Aggregator to "count" for the exact count the API key was used on the selected period or use another aggregator based on your desired results.
Final configuration will look similar to this:
I am trying to get my application(where hosted in google cloud) uptime history to a my own page. Is there any api so something on google cloud? I only need to get date and the up/down percentage or time.
I am already configure the uptime checks on google console. But I need to integrate this into my application.
Yes, you can but it's not obvious and it may be easier to use something other than Cloud Monitoring to export uptime data to a non-GCP site :-)
If you do want to use Cloud Monitoring to source this data into an off-GCP page, one of the Cloud Monitoring SDKs may be best. You can create a URL too (see below) but you'll need to authenticate this URL and that may make it too complex.
By way of an example, here's an Uptime check I created against my blog:
I recommend Google APIs Explorer as it's an excellent way to understand Google's services (via the REST APIs) and to test an approach.
First: List|Get Uptime Check(s)
https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.uptimeCheckConfigs/list
Plug in to the form on the right hand side parent, the value of projects/${PROJECT}
If your Project ID is freddie-210224-66311747 then you'd type project/freddie-210224-66311747.
https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.uptimeCheckConfigs/get
For this one, you need to provide name, the value of projects/${PROJECT}/uptimeCheckConfigs/${UPTIME_CHECK}
If your Uptime check is called test, then you'd type projects/freddie-210224-66311747/uptimeCheckConfigs/test
NOTE In my case, I used an Uptime check name that included periods (my.blog.com) and this was converted (to my-blog-com). So, you may want to list first to check the name.
Click "Execute" (You don't need to have API Key checked but it makes no difference).
What I learned is that Uptime checks are Metrics like all others. I confirmed this by watching the Chrome Dev Tools while I was watching Uptime checks.
Ensure that you use the correct metric name. You can use Monitoring's Metrics Explorer to confirm this:
The Resource Type is Uptime Check URL (uptime_url)
One (!) of the Metrics you may use is Request Latency (monitoring.googleapis.com/uptime_check/request_latency)
If you populate the Metrics Explorer, you should see the same data plotted as with the Uptime Check page.
Click Query Editor to get your Uptime Metric represented as Cloud Monitoring Query Language (MQL), remove any line-feeds. You can use:
fetch uptime_url | metric 'monitoring.googleapis.com/uptime_check/request_latency' | group_by 1m, [value_request_latency_mean: mean(value.request_latency)] | every 1m
So, now we want to query Montioring Metric Time-series
https://cloud.google.com/monitoring/api/ref_v3/rest/v3/projects.timeSeries/query
The value for name is projects/${PROJECT}
For query, paste in the MQL from above retain the quotes, i.e. "fetch uptime_url ..."
Hit EXECUTE
You should receive a snapshot of the time-series data underlying your Uptime URL. You can revise the MQL to reflect exactly the subset that you need. At 2021-02-24T20:55:38 the latency was 20.869:
So, to get e.g. request latencies for your uptime checks, you can use the Monitoring API's TimeSeries Query method and, with a suitable Query, this will yield JSON data including an array of Point (values). These values could then be transformed and surfaced into your external page.
We are working in one of the customer module. With the help of this module, we are calculating price for google cloud components like compute engine with attributes images, boot disk, region, snapshot etc.
But we found that GCP is revising the prices and JSON is not modifying immediately. Newly revised pricing values comes in JSON after few days.
So the price according to JSON and price in google cosole/calculator are different in this case.
Is there any exchangeable format to get revised pricing immediately?
The JSON file isn't exactly the ideal way to obtain pricing information anymore. In fact, the link to the JSON file was removed from the Pricing Calculator page very recently precisely because they want people to start using the Catalogue API. That one has the prices up-to-date for sure, so I'd recommend you to use that instead.
I'm trying to add a new product to my seller account, using SubmitFeed function (feed type being _POST_PRODUCT_DATA_). Everytime I try it, a different error comes up. Want to confirm if I understand the underlying concepts correct.
ISBN/UPC/EAN are standard global identifiers used to identity a commodity uniquely.
ASIN are standard Amazon identifiers used to identify a commodity on Amazon uniquely.
SKU is my personal unique identifier.
So, if I want to sell a product that is existent on Amazon, I can specify ASIN/UPC/EAN/ISBN. What is benefit of providing Description Data as it won't affect the description already showing on Product Listing Page on Amazon
I can add a new product (not existent on Amazon) by specifying my local SKU and omitting any ASIN/UPC/EAN/ISBN. Are there any specific rules for mandatory fields/data to be specified while adding product with specific categories/product-types?
Your understanding of the underlying concepts seems accurate.
There is no direct benefit to providing tons of details to an already existing product. Unless Amazon for some reason switches to "your" data from whoever else's. There are rumors of that happening, but I haven't had such a case myself.
For the most part, you cannot add a new product without specifying a ASIN/UPC/EAN/ISBN. There are exceptions to this, but they are few and far between. It is in Amazon's interest that identical products will be on the same page. If the description or other information on that page is wrong, contact Amazon. Apart from what the XSDs define as mandatory, there are category specific mandatory fields. The easiest way to find them is by creating a product manually in Seller Central.
According to Google Analytics, I had 5 visits from zero unique visitors. Is that a bug or did I perhaps implement something wrongly? Or hasn't the data processing finished yet (I created this view 2 days ago)?
The view is based on an include custom filter that's supposed to include only traffic from any of three ip addresses. The regex I used for this is
62\.58\.32\.193|77\.172\.143\.12$|213\.125\.166\.98
My best guess would be the way Google defines unique Visitors. Sometimes I have been visiting my own website periodically and I ended up showing up as a unique visitor (My site isn't so popular so it's easy for me to track that). I would either have to say that it has to do with the nature of visits or the actual way of unique visitors. According to google this is how the find unique visitors
The other Unique Visitors metric calculation (Calculation #2) is based
on the __utma cookie. Calculation #2 is used when segmenting the
Audience Overview report or when viewing Unique Visitors over any
dimension other than date. As such, Calculation #2 is used in custom
reports to allow for the calculation of Unique Visitors over any
dimension, such as browser, city, or traffic source.
source: https://support.google.com/analytics/answer/2992042?hl=en
Occasionally, there are problems with Google Analytics reporting. Check the product forums. For example, here is an issue that happened on Nov 11, 2013:
http://productforums.google.com/forum/#!topic/analytics/fsurDK8AOcY
This issue can also crop up when you are using the page dimension. Unique visitors are only assigned to the first page in a visit as described here. But, it doesn't seem like that is the case for you.
Finally, its possible, analogous to the page dimension situation, that unique visitors are only assigned to the first IP address that a visitor came from. If that is true, then if the people who came to your site had previously come from a different IP address, then they wouldn't show up as unique visitors in your filter.