I am using the Adobe Analytics connector for one of my reports. One of my 'Dimensions' is called 'Websites' and the data is a URL (www.myurl.com), we look at over 100 websites and I am looking at the traffic for each one. When I use the Websites dimension it only displays 10 of the sites out of the total number. Is there a way I can add a line of code in the advanced query editor to remove this limit? Or is this a limitation with the Power BI connector?
Using the 'Top parameter'(returns up to 50,000 dimensions per call) fixes my problem and breaks down my URL column by the correct number of dimensions as opposed to 10.
{Cube.ApplyParameter, "Top", {10000, "evar123"}},
Related
I need help with improving refresh times on a Power BI dashboard with about 20M rows of data and 80 columns pulling from SQL Server. I cannot use Power BI Service in any capacity, this has to load into Power BI Desktop.
My refresh times on the raw data (virtually no transformations in Power Query) are taking about 3-4 hours.
Microsoft recommends incremental refresh to archive my historical data and only refresh the latest changes, but that requires Service and I 100% cannot use it.
Is there any other way to significantly improve my refresh times beyond Service's incremental refresh? If it was under an hour I'd be happy.
What I've tried:
Native Query to leverage the server
reducing column selections
removing all transformations
Splitting tables in Power Query and selectively turning off refresh in the historical tables - as soon as they get stacked/appended Power Query triggers a refresh on all stacked tables regardless of which ones have refresh turned off.
Looked into Power Query PQFL/M code to activate refresh of tables - can't find any method/property to control this in M code.
optimizing the SQL, haven't gotten any significant improvements.
20 million rows should not take that long, especially with no transformations. Something else is going wrong but without access to your data and hardware, it is impossible to say.
One possibility is do an initial data load and then turn off refresh on that query. Add a new query for just the new data (which should be quick) but load the new query to a completely new table. In PBI, you will then have two tables. Create a calculated table in DAX which is a union of your old, non-refreshed data and your new data. Refreshes should be very quick after your first load but obviously you need to think about how it scales as your data grows.
Good Day
A client I am working with wants to use a PowerBI dashboard to display in their call centre with stats pulled from an Azure SQL Database.
Their specific requirement is that the dashboard automaticly refresh every minute between their operating hours (8am - 5pm).
I have been researching this a bit but can't find a definitive answer.
Is it possible for PowerBI to automaticly refresh every 1min?
Is it dependant on the type of license and/or the type of connection (DIRECTQUERY vs IMPORT)
You can set a report to refresh on a direct query source, using the automatic report refresh feature.
https://learn.microsoft.com/en-us/power-bi/create-reports/desktop-automatic-page-refresh
This will allow you to refresh the report every 1 minute or other defined interval. This is report only, not dashboards as it is desktop only.
When publishing to the service you will be limited to a minimum refresh of 30 mins, unless you have a dedicated capacity. You could add an A1 Power BI Embedded SKU and turn it on then off during business hours to reduce the cost. Which would work out around £200 per month.
Other options for importing data would be to set a Logic App or Power Automate task to refresh the dataset using an API call, for a lower level of frequency, say 5 mins. It would be best to optimise your query to return a small amount of pre aggregated data to the dataset.
You can use Power Automate to schedule refresh your dataset more than 48 times a day. You can refresh it every minute with Power Automate, it looks like. I can also see that you may be able to refresh your dataset more frequently than that with other tools.
Refreshing the data with 1 min frequency is not possible in PowerBI. If you are not using powerBI premium than you can schedule upto 8 times in a day, with the minimum gap of 15 minutes. If in case you are using PowerBI premium you are allowed to schedule 48 slots.
If you are not able to compromise with the above restrictions, then it might be worth to look into powerBI reports for streaming datasets. But then again there are some cons to that as well, like they work only with DirectQuery etc etc.
I have Power BI report which is connected to the dynamic 365 to show the report of contact and account but the data set has more then 2 GB of data and I am not able to publish the report.
how can I decrease the size of the data set so that when publish I can refresh and update data set to get the 2 GB of data.
I am not able to publish the report
one thing which I used to take the top 100 rows by using one of the option in the Data Set but when I refresh the data set after publish it get top 100 record
A .pbix file has no size limit, but PowerBI service has a limit of 1GB (for a pro subscription) per dataset, higher-level subscriptions (embedded/premium) have higher limits.
Unless you can upgrade your subscription (which may not be worth it for one report), you have to decrease the report size.
The suggestions are always the same and can be found on MS website, I had a look at it and I have nothing to add.
Below the points that you will find on the MS website, in case the link stops working:
Remove unnecessary columns
Remove unnecessary rows
Group by and summarize > pre-aggregate/aggregate data, with the detail you need
Optimize column data types > chose the right data type for each column
Preference for custom columns > create custom columns in PowerQuery (M), they have a better compression rate compared to DAX calculated column
Disable Power Query query load > do not load tables you don't need (support table used for calculations but not needed in the model)
Disable auto date/time > disables the calendar hierarchy created by PBI for each date in your model
Switch to Mixed mode > This mode is a mix of import and direct-query, you will find more info online about this. (if you choose this have a look at the aggregations functionality)
How many reports can I create in one power bi workspace ?
How many datasources can I create over there ?
Details:
We are going to implement our own application and use Power BI Embedded features. Our app will own all data, thus our users will be working with reports on behalf of one master Power BI Pro account. Since every user could have multiple reports the question is how many reports can I store into one workspace ? I found some links where developers claim, that only 200 reports can be saved into one workspace, but I've not found any official doc on this. Currently we created one report in workspace and cloned it via saveAs function of Power BI JS SDK. So far I could create more than 1000 reports. So what limits are ?
UPDATE1
So far I reached limit in 1005 reports in one workspace. I can not create new one 1006s report neither from Power BI Services nor API call. Just got error "Unable to save the report" with some technical details. Once I removed one report I can create new one, but again can't exceed limit in 1005 reports.
UPDATE2
So when I save report (in Power BI Services ) got standard message box with error and link to "technical details", when click on it I set of ids:
Using embedded JS SDK got similar thing:
function GetCurrentReport() {
var embedContainer = $('#reportContainer')[0];
return powerbi.get(embedContainer);
}
const report = GetCurrentReport();
report.saveAs({name:"NewReportName1111"});
In both cases 500 error came from some APIs.
UPDATE3
It looks I reach limit, Fidler says:
some API throws ResourceLimitsReportCountExceeded exception.
Do someone have any ideas how to increase this CountPermitted ?
Recently I contacted Microsoft representative and asked all those questions and more.
There are questions and answers this person got, I believe, from Power BI product team:
What is max count of reports per workspace ?
ANSWER: You have a limit of 1000 datasets per WS, or 1000 reports per dataset. You have probably reached one of them.
What is max count of workspaces per one PRO user ?
ANSWER: Right now it is 250, will be increased to 1000 with the new workspace experience that is now in public preview
Is it possible to extend that limit of reports in workspace to some bigger number ?
ANSWER: Not right now
If using dedicated capacity Power BI Embedded App Services , can we manage such limits ?
ANSWER: These limits are not related to using capacity. You must purchase a dedicated capacity to go to production with PBI Embedded.
What version of Power BI API should we use ? Version 1 or Version 2 ?
ANSWER: There is only 1 API version- V1. The ‘V2’ you see is for the SDK of PBI API
Microsoft Power BI provides work-space to create or publish report from power BI desktop. Below are details of data-set and report capacity in Power BI service for free or Pro users.
Free users have a maximum 1 GB data capacity.
Pro users of Power BI Pro have 10 GB maximum capacity.
Pro users can create groups, with a maximum 10 GB data capacity each.
There is no limitation on number of reports but the total size of all the reports should not exceed 1 GB for free user and 10 GB for Pro user. Also, Pro user can create 10 GB work-space for each user.
Cheers.
If you are using a PowerBI pro account, each workspace has a 10GB storage limit. If you want to upgrade to premium, the storage space depends on the plan you choose. For more details see this : Storage in PowerBI
I have a Power BI Desktop application that accesses one Data Source which is an OData feed. When refreshing the data in Power BI Desktop it takes much longer . After clicking on Refresh in the popup it shows "Evaluating" and "Loading 2 KB..." for a very long time. From logs of OData API,I can see that the OData feed is called twice and each request is taking same amount of time.
Why does Power BI desktop call the OData feed multiple times? Is there a way to reduce the number of calls?
Note:
I have gone through Referenced Queries And Caching In Power BI And
Power Query Blog But in my case my not using any transformation.
Just loading OData feed.
I have gone through the post for same
issue at Power BI community and try the suggested approach but
still issue persists.
Case is similar to "why-is-power-bi-running-my-sql-query-twice"
https://blog.crossjoin.co.uk/2020/07/05/why-is-power-bi-running-my-sql-query-twice/
Power BI wants to know the schema of the table before the query actually runs, so it asks Power Query to return the top 0 rows. Probably its also true for OData.
Using M function Table.View should help.
https://blog.crossjoin.co.uk/2020/05/14/speed-up-data-refresh-performance-in-power-bi-desktop-using-table-view/