-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Make page level data available in BigQuery #10
Comments
Unfortunately we're unable to add page-level data to BigQuery. Could you describe the API limitations you're hitting? Also are you doing any kind of rate limiting or batching? |
The official api docs aren't super clear about quotas, but according to https://github.com/treosh/crux-api#batch-request, each individual request inside the batch counts towards the quota. Before today, everything I read seemed to point at
As mentioned here, being able to query for coverage instead of individually hitting the api repeatedly would reduce the need for so much quota. Are you able to share anything about the reasoning for not making page level data in BigQuery? Are there privacy concerns? |
The Treo docs are correct that queries within a batched request still count towards the quota.
Could you elaborate on what you mean by "query for coverage"? Not sure if it's referring to getting feedback on current quota usage or a feature request for better coverage of URLs.
Yeah, we would want to avoid anyone being able to say "show me all pages for a given origin" even if it's not their site. Site owners should know what all of their URLs are and how popular they are, so it should be possible to create an ordered list of URLs and to query the most popular ones, which are most likely to be included in the dataset and have the biggest influence over the site's aggregate CWV performance. |
The origin level data is already there, and BigQuery is perfectly suited for broader page level analysis. The api quotas make it hard to analyze large sites.
The text was updated successfully, but these errors were encountered: