We have a requirement where we have to fetch data from Power BI dataset and transform the data and then produce the data again in Power BI Workspace so that I may consume that in PBI.
-> Is there any way I may connect my Power BI Dataset in ADF.
-> Once the data is fetched can I use Power Query to work on dataset in ADF.
-> Where to store that data: in ADLS or in Azure BLOB or may I save this data in SQL directly?
-> Once I have transformed the data can i publish my gold data in Power BI Workspace?
There is PowerBI REST API to execute queries:
Datasets – Execute Queries
you can access the power bi dataset via managed identity or service principal authentication.
The below blog :
https://datasharkx.wordpress.com/2022/11/03/copy-data-from-power-bi-through-azure-data-factory-synapse-pipeline-via-managed-identity-authentication-part-4/
You can leverage Dataflows to do the transformation
The query w.r.t 3rd aspect really depends on your architecture.You can preserve it anywhere
Related
Hi I am Consuming Snowflake View in Power BI Dashboard ; I noticed that Power BI reporting tool design is to run the Snowflake view (Direct Query Method)and consume all the data coming out of Snowflake View and then do the further filtering, transforming on the data in its own memory place. This has performance repercussions.
My Question is how do i bring Power BI User Selected Parameters filtered data from Snowflake View into PowerBI ; as far as i know Snowflake View doesn't have Parameterized View. Can i consume Snowflake UDTF (User Defined Table Function )in Power BI, i understand that Snowflake UDTF object can be Parameterized.
You can use QueryFolding (predicate push-down) in PowerBI. This will push the transformation to the source.
You will have to check the PowerBI documentation to see how this is done properly.
It should speed up all queries.
Snowflake
PowerBI
Is it possible to use different warehouse in snowflake for data refresh in Power BI than warehouse used for dashboard development?
I'd like to have separate warehouse for data refresh in order to be able to track snowflake expenses.
Let's say I'm using SF_TUTS_WH as a warehouse for development, but I'd like to use PBI_WH as warehouse for data refresh in PowerBI. Is it possible to achieve this or do I have to change snowflake warehouse in PowerBI file before I'll load it to PBI online?
FYI: https://www.snowflake.com/blog/using-sso-between-power-bi-and-snowflake/
One option which will work for you is to have the ware house name paramterized for the reports. The value of the parameter can be updated from the service based on the requirements.Please check out the below video link to see how the parameters can be updated in the power bi service.
https://youtu.be/MJVSu47iMmo
Hope this helps
Thanks ,
Prasanth
We were implementing the Azure Time Series insights(TSI) for the data streaming through IoT hub and wanting to connect to Power BI for advanced visualization. There is a provision to build a custom query to connect to Power BI from Azure TSI but i guess there is a limitation of fetching a dataset for specific time frame and only for the time series ID selected for the view.
Is there a way we can get the latest data in Power BI as and when the data refreshes in Azure TSI and for additional number of time series ID which were not selected while building the query.
Cheers,
Amit
If you generate a query from TSI to feed into Power BI, you can manually edit it to include the TSIDs you want to see. If you set the "isSearchSpanRelative" parameter to true, the data should refresh in Power BI as it comes into TSI.
I have a collection of .pbix models that follow a similar structure, ie, have the same tables and relationships.
It is too complex to combine them all into a single .pbix.
Is there a way to upload all these tables into a single repository, like PBI Service dataflows or a data warehouse, or something similar.
And then get the data back to PBI Desktop and perform DAX calculations, visualizations and report.
Any suggestions/ ideas?
Thank you so much for helping!
You can publish them to Power BI Service, and then create separate reports, but using these published datasets as a data source.
See Connect to datasets in the Power BI service from Power BI Desktop.
After publishing your "model" reports to Power BI Online, start making a new blank report, but instead of getting the data from files/databases/etc., choose Power BI service as a data source and select the previously published dataset. After that, you can publish your report the same way, but in this case you can share one dataset (your model) between multiple reports.
I am connecting the Power BI to Azure Data Lake store with multiple files representing multiple tables.
1)
update It is currently loading the data into the Power BI file.update
But, can I have live connection from Power BI Desktop with the Azure Data Lake Store?
2)
Can I load multiple files to represent Dimensions and Fact tables.
From desktop you can access the Azure Data Lake Store data source - just make sure you're using a recent version of Power BI Desktop Data Lake Store - Power BI.
You can join multiple queries together in Power BI desktop
The documentation for PowerBI does not list Azure Data Lake Store as a source that can be connected live as of May 2018: https://learn.microsoft.com/en-us/power-bi/refresh-data#live-connections-and-directquery-to-on-premises-data-sources.
Alternatively, you could try using Azure Stream Analytics to create a job that can copy data, and connect to the live stream, but that process might need to be manually triggered and requires data movement, which might not be ideal for your scenario. https://learn.microsoft.com/en-us/azure/data-lake-store/data-lake-store-stream-analytics
Hope this helps.
Yes, you can get data from Azure Data Lake Store from Power BI desktop application:
Also, you can join multiple queries from different tables into PBI desktop app.